Full-stack observability with actually useful AI | Grafana Cloud
Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.
Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
Create free account
$300 in Free Credit Towards Top Cloud Services
Build VMs, containers, AI, databases, storage—all in one place.
Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
The DotNetWikiBot Framework was developed so that it can offer a helping hand with many complicated and routine tasks of wiki site development and maintenance. DotNetWikiBot Framework is a cross-platform full-featured client API, that allows you to build programs and web robots easily to manage information on MediaWiki-powered sites. DotNetWikiBot Framework can also be used for learning C# and .NET.
creole/c is a Wiki Creole parser and a HTML converter. It implements Wiki Creole 1.0 and almost all of its additions. The parser is written in C++ and has a simple event driven plain C API. The converter is a stand-alone console application.
Un ensemble de scripts utilises par Framasoft. Ils ont ete realise afin de mieux gerer des serveurs ou pour ajouter des fonctionnalites à nos sites internet.
A Python library and collection of tools that automate work on MediaWiki sites. Originally designed for Wikipedia, it is now used throughout the Wikimedia Foundation's projects and on many other MediaWiki wikis including wikidata service.
We do not use sourceforge.net anymore, but are very grateful for their support in the past. Please see our website at https://www.mediawiki.org/wiki/Pywikibot
Our pypi package could be found at https://pypi.org/project/pywikibot/
A collection of python scripts to create and handle an XML corpus (a large collection of text for linguistic purpose) from an original Wikipedia database backup dump. It includes a regular expression based parser for the MediaWiki markup language.