Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 8 years ago.
Improve this question
I'm using a stream engine that updates my database every second with significant set of data.
I would suggest looking into using a WCF Duplex Service. I found several articles about implementing it searching for "Silverlight WCF Duplex Service." You should also consider bring back chunks of your data instead of the whole set each time if the data set is significantly large as you mentioned.
Related
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 3 years ago.
Improve this question
How to integrate Apache Flink with Timescale Db.Explain How integration is done along with some example codes.The problem is Kafka writes into a topic flink reads it and after some calculation it should write in Timescale Db.
You should be able to use Flink's Table API with the JDBC connector with the PostgreSQL driver to write to TimescaleDB. If you get stuck trying to make that work, ask a more specific question, showing what you've tried and explain what's going wrong.
Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 6 years ago.
Improve this question
In the context of RESTful APIs, what is an example of a real world object, that would make sense to Create but not Lookup? Is there one?
--Edit--
The question was to ask if there exists a usecase where a Create API would make sense, but not necessarily a Lookup API for the object - in places where you wouldn't want to expose the data(using the lookup API) to the outside world.
How about a keystroke logger...sending POSTs to create the audit trail.
You would not want to allow anyone to GET the results (besides you).
Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 8 years ago.
Improve this question
I'm currently using cf 10 and making alot of use using cfsearch (solr) and I'm just wondering if there is a way to report / record what people are searching for on my website.
I have read that this is possible when using solr directly on other platforms... So I'm hoping there is an easy way to retrieve this via a cf tag
Any ideas greatly appreciated.
Assuming you control the form that is used to submit the search parameters, you can simply grab the contents of the search field(s) (before or after you have completed your search - or as Dan pointed out, you could spin that process off into a cfthread to be done concurrently) and store that information however you see fit.
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 9 years ago.
Improve this question
I don't know if this question is too simplistic - but are there any sort of best practices or guidelines when it comes to deciding how much data is too much for client side processing (sorting/filtering) with AngularJS.
I am wondering if it make sense to build some sort of trigger into my code - perhaps when the data set reaches a certain size do all manipulation on the server side, if its below that size do it on client side. Is that overkill? Am I over thinking this?
Thanks for your feedback!
You should only every send the amount of data to the client that it needs, that includes aggregate data if the client only needs aggregate data.
The most to send is also however much the client need to function.
Optimise for performance AFTER you have a working client, not before.
This is true for any client, not just Angular.
Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 8 years ago.
Improve this question
I have a lot of data types - datetime, e-mail, data - for sending notification.
If I use the normal database, I need to poll the database by surveying every second.
Please, help me with choice of architecture.
I want store this data in RAM - maybe redis or memcashedb
rabbitmq - is open source message broker software (i.e., message-oriented middleware) that implements the Advanced Message Queuing Protocol (AMQP) standard