Should webgl be used for simple websites? [closed] - webgl

Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 4 years ago.
Improve this question
Should webgl be used for simple websites?
I'm not sure or it is wise to use webgl for a simple website just to give it a better look. Will this work on all devices?

WebGL is widely supported today https://www.caniuse.com/#feat=webgl
Whether you "should" use it or not is a broad question. Remember that you aim at improving the user experience. People are forgiving when they play video games, but they don't want to hear their computer fans spin, witness their battery discharging very fast or feel their device getting hot when all they wanted was to read a cooking recipe. Try to be user friendly.
You may for instance want to cap the framerate and/or reduce the resolution on high definition devices, pause the animation loop when the window looses focus (which is not the default behaviour of requestAnimationFrame) or when there is nothing changing on the screen (if the WebGL element is interactive for example). Also, try to write efficient algorithms: it's easy to start writing things on the fragment shader or the CPU when they should be done on the vertex shader. There are many ways to accomplish the same thing and they don't put the same stress on the computer.

Related

design pattern to use utility app connecting to internet [closed]

Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 7 years ago.
Improve this question
This may be a generic question but still i need suggestion/guidance on which design pattern or architecture to follow for developing this app.
The app is for premium users(for iPhone) on airport and will be able to access the internet at higher speed and on multiple devices simultaneously.
I'm planning to go for Singleton,Factory,MVC patterns and client-server architecture for the app.
What-all things do i need to re-think and then design the app or are above patterns sufficient to go ahead ?
thanks
It seems bizarre to come up with a list of patterns to use before you look at what your app is actually going to do. This is not how you use design patterns.
The way to use patterns: When you have a problem, and you think about how to solve it, you try to find out whether your problem fits a well-known pattern and adapt that pattern to your problem. Or you figure out that it doesn't fit any well-known pattern, and then you solve the problem without using any pattern.
This is like going to a shop buying blue and yellow paint, and then deciding what parts of your home you want to paint. You do it the other way round. You decide what needs painting, then you decide what would be a nice colour, then you buy the colour. You don't buy the paint first. You don't decide on design patterns first.

Robotics Project based on slam algorithm [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 7 years ago.
Improve this question
I am very beginner in robotics. I want to make a robotics project based on slam algorithms. I know many algorithm and i have the confidence to implement it in any language but i dont have any idea based on image processing and hardware. So, can anyone give a tuotorial based on slam based robotics projects[including how hardware organized and how image processing is done for that project], after seeing that i can make a slam based robotics project from my own.
In addition, If anyone give me a video lecture series for that then it would be very helpful.
Thanks in advance.
I have tried to do something similar last year. I created two systems. The first system made use of a camera and laser to detect objects and determine their location relative to the system itself. The second system was a little robot with tracks (wheels would be better), that used dead reckoning to keep track of its own location relative to its starting location. The techniques worked really well, but unfortunately I did not have the time to combine the two systems. I can however provide you with some documentation that was incredibly useful for me at that time.
These tutorials provide information on both the hardware and the software.
Optical Triangulation (detection of objects with a camera and laser) :
http://www.seattlerobotics.org/encoder/200110/vision.htm
Dead Reckoning (a technique to keep track of one's own location) :
http://www.seattlerobotics.org/encoder/200010/dead_reckoning_article.html

Image Recognition in iOS [closed]

Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 9 years ago.
Improve this question
I am creating an iPad app for golfers, I will get their score card image as below. I want to calculate sum of scores for each person by scanning this image.
Is there any way. Please give me a logic for this.
I have used opencv http://opencv.org in the past to do something similar but with sudoku puzzles.
It is a LOT of work and making it work with handwriting will add to the difficulty.
I found a really good resource for analysing sudoku grids. I'll try to find it again but it was 4 years ago.
Good luck though.
There is a Tesseract port for iOS which is about the best OCR you're likely to get on the platform without either:
A) Porting another OCR library
or
B) Shipping the images off to an online OCR service
To make this more complex, you don't just want to OCR but you want to OCR handwriting and put it into a grid. This is not something that can be done overnight but is in fact rather complex. Not impossible, but complex.
Would it not just be easier to let the players enter there scores straight onto the app and then airprint a score card?

How would I go about making a pixel based game? [closed]

Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 9 years ago.
Improve this question
Something like galaga! I have no idea where to start. I know how to program, I have an environment to make the game, and I can create all the pixel/tiles on the screen. I just don't really know where to go after that. I've looked online for help but with no luck.
Any feedback is much appreciated!
A great alternative would be to use Löve2D, it is a game framework that uses Lua as its programming language and it is very easy to use.
No matter what language you code in, I would start by making a diagram describing the game flow. If you break it into small pieces, you'll find yourself following that process as you start to code.
Ask yourself:
What is happening in the game? (Aliens attacking a single ship)
What is the object of the game? (Destroy each successive wave of aliens without getting nuked.)
What controls are necessary to accomplish this? (firing button, movements, any specials you might incorporate.)
what sorts of power-ups / extra lives will be available and when?
What is the reward schema (more points / power-ups for higher-level kills, etc.)
What is the derivative increase in difficulty? (is it linear - gets a bit harder each level or, does it get increasingly difficult by greater amounts as you pass levels - also, do you want to build in respite levels where it chills for a level or two.)
Hopefully, you get the idea - just chop up the project into bite-sized chunks before you write your first line, follow your process, and adjust as necessary - a couple of hours of planning could save you many hours of dithering while trying to code a project you have not really defined for yourself.
I hope this helps
~b

Is there a good library for cutting out people in a still photo? [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 1 year ago.
Improve this question
Are there any libraries, in any language, out there to help identify and grab the images of people in a still photo? Something similar in effect to the way the Kinect can isolate users.
Thanks much!
I think it depends very much on the setup (e.g. simple bg. with decent lighting condition vs. random bg. with random lighting). If you can make life easier for yourself and isolate a few simpler use cases that would be great. Still there are other available method, look at the plethora of research around pedestrian detection for example.
One thing I did try and it works surprisingly well although computationally intensive is the Histogram of Gradient Orientations, implemented in OpenCV as the HoG descriptor. For a still photo this should produce decent results. You can have a look at the OpenCV sample. I also recommend having a look at Dramanan's excellent papers.
Long story short, thanks for years of inspiring research in computer vision, there are quite a few interesting options out there, it's up to how willing you are to go into detail. Still, regardless of how clever algorithms can be, I believe it's far more important to get a decent setup that allows simple and efficient solutions rather than complex solutions that try to cater for every possible situation. Goodluck!

Resources