What are the plans for the future of the REST plugin? Has anybody tried to make it more JSR-311 like? I am not a REST user myself, but I am kind of bored and could help if there was a clear understanding of what needs to be done.
musachy On Mon, Aug 25, 2008 at 12:30 AM, Jeromy Evans <[EMAIL PROTECTED]> wrote: > Don Brown wrote: >> >> On Mon, Aug 25, 2008 at 12:54 PM, Martin Cooper <[EMAIL PROTECTED]> >> wrote: >> >>> >>> Another option is a client-side component-based framework like Ext or >>> Flex >>> running directly against web services, RESTful or otherwise. No >>> server-side >>> web framework required. Of course, you could use something server-side >>> like >>> DWR to facilitate working with web services, or Jersey for RESTful >>> services, >>> but that would be a choice rather than a requirement. >>> >> >> This is a nice design, when you can do it. GWT is also a good way to >> build these types of apps. Unfortunately, they can easily break much >> of what makes the web what it is - the back button, unique, >> addressable URI's, accessibility, search engine crawling, etc. >> Therefore, I think some sort of server-side web framework will usually >> be necessary, however, I don't think it has to go so far as JSF, where >> they try to push all the state to the server. I was talking with a >> guy here at work who is looking to start using GWT more about how and >> where a plain HTML view of the application fits. He wants to do very >> dynamic, client-side heavy views, but still needs to support search >> engines and REST clients. What if you use Jersey for your REST API, >> GWT or straight JQuery for your client-side UI, then have Jersey + >> something generate HTML views of your REST API, which you could use >> for search engines and developers wanting to browse and interact with >> your application. If you can have the HTML representation of your >> REST API auto-generated, you wouldn't have to maintain two different >> interfaces, and you could go fully nuts with your client-side heavy >> app without having to worry about accessibility or search engine >> issues. >> >> Don >> >> > > > [rant] Personally I think search engines need to solve this problem. The > era of crawling sites needs to close. As a publisher of content I should be > able to connect to a Google API and publish my content and URIs to them in a > standard machine-friendly format ready for indexing. Alternatively, I could > implement a dedicated-service for them to consume instead of emulating pages > of content in a non-page-oriented application. Then my application then can > be what it needs to be in any form suitable for my users instead of > perpetuating the artificial SEO-optimzation industry. [/rant] > > Anyway, despite that, I took this approach recently with a client-heavy > (single page) application myself, with the exception of autogeneration of > the HTML. Basically: > - mandated that the client include a custom header (X-RequestedBy) and > signature in the request > - if headers present, the S2 rest plugin handled the request and returned > the resource in the requested content type. I just had to build the view > myself for html. > - if the header's not present and it was a GET, the REST plugin returned the > HTML view and sitemesh decorated it as a full HTML PAGE. > - if a resource was requested directly and the user had javascript, they > were redirected to the rich client with the best-guess initial state based > on the URI > - all flow control is managed on the client. > > That meant that one action could service requests for the resource for rich > clients and support search engines requests for the same content. > Search engines could browse the site through the same content spread over > many little well-formed pages. > Users accessing the site via the search engine's sub-URI would see the rich > client with appropriate initial state derived from the URI > On the client-side sensible URIs could still be used in links and listeners > adjusted the content type when appropriate. > > Users without JS could get by but were a low priority. Users with screen > readers are still a challenge but not due to struts. > > This approach wasn't as simple as it should be though but confirms that > Don's idea is feasible. The biggest problem was in fact with IE6 memory > leaks and the poor performance of javascript in most browsers. A flex > client could have used the same services without a problem. If automation > of a bland html view with a sitemap were provided for users without > javascript/flash you'd eliminate the double-up on the views for search > engines. > > I definitely like the direction these discussions are going. > > > > > --------------------------------------------------------------------- > To unsubscribe, e-mail: [EMAIL PROTECTED] > For additional commands, e-mail: [EMAIL PROTECTED] > > -- "Hey you! Would you help me to carry the stone?" Pink Floyd --------------------------------------------------------------------- To unsubscribe, e-mail: [EMAIL PROTECTED] For additional commands, e-mail: [EMAIL PROTECTED]