-
Notifications
You must be signed in to change notification settings - Fork 654
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Distibuted compute #2711
Comments
But what will you save with this? My devices have an interval of 2 minutes which is already far more what effectively is needed. A goal of this project is to make usage as easy as possible for non-techy people. Having to learn about docker is way beyond most users will want to look into. BTW, @jomjol originally had offline processing in docker but moved it all to the ESP. |
For me running at 240mhz each processing takes 3 minutes with analysis of 4 digits and 4 analog clocks. I would like to speed this up since 3-4 minutes delay equals to around 70-90 liters of hot water when both kids shower. If we could distribute compute between 2-4 esp devices it would be 1,5 to less than a minute, resulting in better resolution and the possibility to take action before hot water is all spent. Power management on an esp not running on battery is not really an issue |
@stefanh12: If you'd like to speed up the process it is possible to optimize this without adding addtional computing power / complexity. The following things affect the processing time.
Just for reference:
|
@Slider0007: Would be in my opinion good to add this in the wiki. So if someone has needs to speed up, he can find this parameter setting szenario quickly. Edit: I have not done any of the recommendations above and my Round Completed time is in average 35 sec |
That helped a lot, processing time is now way lower! |
Do improve it further you can set the CPU frequency to 240MHz |
We are always open for support. I am also willing to show how to extend the documentation! I have a device which processes 3 digits and 4 pointers, taking 50s at 160 MHz |
Ok, let me know how it works. |
can you have a look on https://github.com/jomjol/AI-on-the-edge-device-docs/blob/main/README.md for an introduction? |
Ok, mache ich... |
The Feature
I've been using AI-on-the-edge-device for a year now and its really great! It would be great if we could offload the processing of the image since it takes quite a long time to either multiple esp32 devices or to a docker container on HA or such. multiple esp32 devices could then split the digital / analog up and really speed up the handling. On HA as a container it would I guess only take seconds to process too.
The text was updated successfully, but these errors were encountered: