Showing 3 open source projects for "network access control"

View related business solutions
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 1

    Empact Foundation Class Library

    Cross-platform C++ library for use as a default application framework.

    A mature cross-platform C++ library for use as a default application framework. Features include: * Threading & synchronization * Socket programming: SSL, NanoMsg & ZMQ * File I/O utilities: zlib, ini, yaml * Native Database access: MySQL, SQLite, BerkleyDB, Postgre, REDIS and ODBC * Built-in mini XML parser; optional EXPAT, LIBXML and MSXML support * Network protocol stack: HTTP, FTP, SMTP, POP3, SOAP, XMLRPC * Scripting languages: Perl, Python, JavaScript, VBScript, Java, Lua, TCL, Squirrel * Cloud Computing: AWS * Encryption: OpenSSL * Platforms: Linux/Posix, Windows, Arduino * Over 500+ highly reusable classes. 4000+ fully documented functions. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    PyGObject for Windows

    PyGObject for Windows

    All-In-One PyGI/PyGObject for Windows Installer

    Cross-platform python dynamic bindings of GObject-based libraries for Windows 32-bit and 64-bit.
    Downloads: 14 This Week
    Last Update:
    See Project
  • 3
    char-rnn

    char-rnn

    Multi-layer Recurrent Neural Networks (LSTM, GRU, RNN)

    ...It supports common recurrent architectures including vanilla RNNs as well as LSTM and GRU variants, letting users compare behavior and output quality across model types. It is straightforward: you provide a single text file, train the model to minimize next-character prediction loss, then sample from the trained network to generate new text one character at a time in the style of the dataset. The project is designed for experimentation, offering tunable settings for depth, hidden size, dropout, sequence length, and sampling temperature to control creativity and coherence. It is frequently used as a learning project for understanding sequence modeling, recurrent training dynamics, and the practical details of text generation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB