I am searching for a component to:
analyzing and comparing sound data
drawing the sound histogram.
Are there any guides or other suggestions on how to accomplish this?
You can check out the wavesurfer program:
WaveSurfer is an Open Source tool for
sound visualization and manipulation.
It has been designed to suit both
novice and advanced users. WaveSurfer
has a simple and logical user
interface that provides functionality
in an intuitive way and which can be
adapted to different tasks. It can be
used as a stand-alone tool for a wide
range of tasks in speech research and
education. Typical applications are
speech/sound analysis and sound
annotation/transcription. WaveSurfer
can also serve as a platform for more
advanced/specialized applications.
This is accomplished either through
extending the WaveSurfer application
with new custom plug-ins or by
embedding WaveSurfer visualization
components in other applications.
It can perform many different types of analysis, I have only used it for practicing Chinese tone pronunciation.
Related
I have read the documentation for the Roboforth environment from STrobotics and recognized that this a nice way for programming a robot. What I missed is a sophisticated software library with predefined motion primitives. For example, for picking up a object, for regrasping or for changing a tool.
In other programming languages like Python or C++, a library is a convenient way for programming repetitive tasks and for storing expert knowledge into machine-readable files. Also a library is good way for not-so-talented programmers to get access on higher-level-functions. In my opinion Forth is the perfect language for implementing such an API, but I didn't find information about it. Where should I search? Are there any examples out there?
I am author of RoboForth, and you make a good point. I have approached the problem of starting off new users with videos on YouTube; see How to... (playlist with 6 items, e.g "ST Robotics How-to number 1 - getting started") which is a playlist covering basics and indeed tool changing.
I never wrote any starter programs, because the physical positions (coordinates) would be different from one user to the next, however I think it can be done, and I will do it. Thanks for the heads up.
I am newbie with drones. I would like to develop a program to manage a drone using opencv to fly indoor over a line.
I am searching a lot of languages but most all of them are GPS based. I saw there is an alternative which calls SLAM to detect the position using the sensors.
Well I have a line in the floor and a camera on my drone. I like mission planner but I am not quite sure if it is the best choice. I will be using Parrot AR, but I would like to use any drone.
So I would like to use mission planner but I am not sure if it is the best choice.
What would be the best SDK you would recommend me to use in order to manage the drone not using the GPS points but relative locations or SLAM?
Well, you have the Parrot API ,and a couple of wrappers in different languages. Node-AreDrone for nodeJs, PyArdrone for python, and there is a wrapper coded in C# which I have used AR.Drone. It has a good user interface which you can see the both cameras, record and replay the videos, control the drone by clicking on buttons, you can see the metrics and configuration of the drone and you have also a way to send commands in a queue. Because I love c# and the features I've mentioned you have already in a user interface, I prefer this. Most of them are quite the same as they use the Parrot API inside by sending udp messages. I couldn't try others, so, there are a lot, and anybody could tell me which one is the best. For mission planner I couldn't find a good solution for indoors. So, for anyone who is lost and do not know here to start as I was. I recommend to select the language you want and search for the corresponding wrapper. If you like c# as me, so AR.Drone is a good choice.
Also if you want to do something with OpenCV. Copterface is a good example. You could implement it in any language with OpenCV.
I'm working on an applicaion in Swift and I was thinking about a way to get Non-Speech sound recognition in my project.
I mean is there a way in which I can take in sound inputs and match them against some predefined sounds already incorporated in the project and if a match occurs, it should do some particular action?
Is there any way to do the above? I'm thinking breaking up the sounds and doing the checks, but can't seem to get any further than that.
My personal experience follows matt's comment above: requires serious technical knowledge.
There are several ways to do this, and one is typically as follows: extract some properties from the sound segment of interest (audio feature extraction), and classify this audio feature vector with some kind of machine learning technique. This typically requires some training phase where the machine learning technique was given some examples to learn what sounds you want to recognize (your predefined sounds) so that it can build a model from that data.
Without knowing what types of sounds you're aiming for to be recognized, maybe our C/C++ SDK available here might do the trick for you: http://www.samplesumo.com/percussive-sound-recognition
There's a technical demo on that page that you can download and try with your sounds. It's a C/C++ library, and there is a Mac, Windows and iOS version, so you should be able to integrate it with a Swift app on iOS. Maybe this will allow you to do what you need?
If you want to develop your own technology, you may want to start by finding and reading some scientific papers using the keywords "sound classification", "audio recognition", "machine listening", "audio feature classification", ...
Matt,
We've been developing a bunch of cool tools to speed up iOS development, specially in Swift. One of these tools is what we called TLSphinx: a Swift wrapper around Pocketsphinx which can perform speech recognition without the audio leaving the device.
I assume TLSphinx can help you solve your problem since it is a totally open source library. Search for it on Github ('TLSphinx') and you can also download our iOS app ('Tryolabs Mobile Showcase') and try the module live to see how it works.
Hope it is useful!
Best!
Since there are lot of high level APIs , libraries and Frameworks available for webGL for developing 3D web applications, I want to select the best (sorry this is bit straightforward) to implement a particular model (which isn't Game oriented) on web. I'm confused how to approach for my work, The criteria I want to use for evaluation is:
pickable objects, easily defined geometry and corresponding texture, multi-camera rendering, possible to incorporate GSLS implementations, type of buffers available.
I can't experiment and judge myself a framework by developing every single demo application in every framework due to time constraint. Is there any particular way to read the documentation for available APIs which mention all these. Moreover, the problem is every framework says they are good in some part and how to overcome this to justify single framework among all those available in net world.
A suggestion would suffice my research...
If you have Maya at hand then www.inka3d.com is easy in terms of defining geometry and texture (because you do it with Maya and your favorite image editor) and you get pickable objects. For shaders you can't use glsl but have to use Maya's node based shader editor.
Is it possible to use the MS Ink API to analyze a scanned image of handwriting, or does it only work with tablet pen input?
If it is possible, where is the relevant documentation?
No, its not possible.
Windows Ink is technology that uses information from the writing process, such as pen up and down, as well as direction of movement. this makes handwriting recognition relatively easy, but only because it can get the live writing data.
Analyzing previously written handwriting is very different and much more difficult, and requires machine learning. In this case it is kind of using the same method of recognition as humans use. check out the developments in Intelligent Character Recognition and Intelligent Word Recognition. It requires a lot of processing power and thats why services such as google goggles have to send the image to a 'trained' machine, and even then it cant really read handwriting well. its cutting edge technology, not ready at all for mass deployment.
http://msdn.microsoft.com/en-us/library/windows/desktop/ms698159(v=VS.85).aspx
This application demonstrates how you can build a handwriting recognition application.The Windows Vista SDK provides versions of this sample in C# and Visual Basic .NET, as well. This topic refers to the Visual Basic .NET sample, but the concepts are the same between versions.
And you must also investigate the main class InkAnalyzer.
http://msdn.microsoft.com/en-us/library/windows/desktop/bb969147(v=VS.85).aspx
This sample shows how to use Ink Analysis to create a form-filling application, where the form is based on a scanned paper form.