Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 9 years ago.
Improve this question
As I understand, SPA means
Move maximum of logic to the client. Means client only receive JSON objects from the server by requesting them and nothing else
Push all our html to as single page
Maybe something else?
Questions
What are reasons/benfits to write Single Page Application?
Especially, what are the benefits to push all our html to a single page?
One of the big differences with SPAs is responsiveness. They just appear more responsive to the user because there's no gaps while a page has to load. Users can also be given direct feedback while they're waiting for changes (like spinning icons or progress bars). This kind of feedback is well received by users and decreases the perception of wait-time. Users can also be interacting with the app in multiple ways, all without having to wait for a new page load for each interaction.
Related
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 7 years ago.
Improve this question
I wish to get the contents of a web page that requires me to be logged in (and one that I do not have control over: e.g. Twitter or Facebook), for example I can have Chrome running and I can see Ajax updating the page updating, but I want to periodically get the contents of this page and somehow save it. I don't mind leaving a computer running to achieve this...
You can use any http software to achieve this (like curl). Depending on the site it will take some investigation of how requests are made, in what order, the post data, the encryption, the user agent, cookies, headers, etc. etc.
It could take some time to find the right recipe.
Generally these sites don't want you to do this though, so don't be surprised when you run up against captcha or other clever methods from preventing exactly what you're trying to do.
Chances are, if you have to ask, you won't get in. But have fun.
Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 7 years ago.
Improve this question
I have created a webpage that takes a skill survey of students of a particular college but they are not ready to give me server yet they would want to see the prototype first how to show them the prototype without using the server but just using the web page that I have created?
A prototype shows how it works without actually working. So instead of getting real data, it can show fake data.
Your survey can show just two or three questions, and these questions can be hard-coded or loaded from a mock object, instead of loaded from the server. Also, instead of actually saving anything, it can just jump to the next question and show a text saying 'Thanks' when you are done.
That way, you can see if your survey has the right look and feel, and they can see that too. Also, it may make it a bit more clear to the server-builders what kind of implementation they need to make. A visible, 'clickable' prototype makes it easier to discuss details in the implementation.
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 8 years ago.
Improve this question
If you're not familiar with the concept of a Microgame, check out this video of WarioWare Twisted.
I'm interested in setting up a site where users can play series of browser-based Microgames which are delivered to them by a server. Ideally this would allow me to crowdsource the games and have an open submission system. What sort of scheme could I use to make this work?
I'm thinking that one way to do it would be to have each game consist of:
A javascript file that defines a MicroGame object that controls a rectangular portion of the screen, gets input and timing information from the main page, then calls back to the main page with a "Success" or "Failure" message.
A folder of assets that must be downloaded before the game executes.
Is this possible to do, client-side within a browser? Where would be a good place to start figuring this out?
There are a lot of open issues here. The biggest problem is what language do they submit games in which you can execute safely on the players machines? That said, there are tools like this out there. You could look at the excellent Play My Code for inspiration.
Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 8 years ago.
Improve this question
For javascript functions that only apply to one page, in a large(ish) web app - is it better to have the scripts inline or in the main javascript file?
If your application gets little traffic, use whatever you like best.
If your application gets a huge amount of traffic, you want to minimize:
the number of distinct requestable resources in your application to maximize cacheability
the size of each individual resource to optimize the empty cache experience.
1 and 2 contradict each other, so you have to pick a tradeoff appropriate to your application.
I would prefer having a proper module system and useful compilation tools so I don't need to have to worry sacrificing efficiency when it comes to keeping my code organized.
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 7 years ago.
Improve this question
I'm attempting to scrape a site that has a split second javascript delay.
I'm currently using python for scraping. Whenever I 'get' the page, the javascript delay has not finished and is has not completely loaded the new dom yet.
How would I scrape such a pge?
You can extend Mozilla to build a web scraper which can leverage the full power of the web browser. After all data have been loaded and the DOM has been built, you can extract needed data from the DOM using XSLT. If the DOM was dynamically changed after initial loading, you can take some approaches to wait for the changes. Visit http://www.gooseeker.com for more information. GooSeeker publish a similiar tool free for everyone. Most of codes are in javascript and readible, from which you can find how it runs.