Everything is ready, however the data is only up to the current point. For the system to operate continuously and always update with the latest data, we need services to schedule and trigger
daily_chplay_app_details
0 6 * * *
: runs at 6 AM every dayNote: Flexible time window is the delay time when running jobs. If you select 5 minutes, the job will start 5 minutes later.
crawl-app-details-maket-chplay
1-role-d4jxk5zk
weekly_chplay_app_reviews
15 6 */7 * *
: runs at 6:15 AM every Sunday of each weekcrawl-review-maket-chplay
1-role-d4jxk5zk
⇒ Do the same for the trigger lambda job processing from the sentiment model
With EventBridge, we have successfully set up periodic schedules for both app_details and app_reviews, and can similarly extend to other jobs like sentiment analysis. This scheduling helps the system:
Automate the data crawling process without manual intervention.
Ensure continuous updates with the latest data, maintaining accuracy and timeliness.
Easily scale to many other tasks in the data pipeline.
Thanks to this, the entire data processing flow becomes proactive, stable, and capable of long-term operation. This is one of the important steps to build a more automated and intelligent data analysis system.