Full-stack observability with actually useful AI | Grafana Cloud
Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.
Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
Create free account
MongoDB Atlas runs apps anywhere
Deploy in 115+ regions with the modern database for every enterprise.
MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
DomainHealth is an open source "zero-config" monitoring tool for WebLogic. It collects important server metrics over time, archives these into CSV files and provides a simple web interface for viewing graphs of current and historical statistics.
Blunder is an automated tool for analyzing chained exceptions in Java. It's usefull for classify, generate a customized error message and a list for possible solutions.
MbeanStat provides a very easy way to instrument Runtime MBeans in WebLogic servers and capture system runtime data like CPU utilization. It also provides scripts to generate graphs for easy analysis. The tool is easy to setup and use.
The application allows the online pdf report generation and the break of a report through one or more dimensions: production or cost reports can output thousands of pages, while a user needs just his own portion of the data.
A Lua-based crawling scripting language and leveraging selenium
I needed a way to crawl a site, crawling using commands. I would put commands in a file or DB to use selenium to interpret the HTML and Javascript. The best would be to have a complete language with conditionals and looping. I'm a java developper and I needed that the crawler to run in a Spring-Boot application. So I decided to use a Lua interpreter in Java to build a crawling tool based on Selenium.
The trick here is to add the crawling commands into the Lua interpreter.