What Integrates with Kilo Code?

Find out what Kilo Code integrations exist in 2025. Learn what software and services currently integrate with Kilo Code, and sort them by reviews, cost, features, and more. Below is a list of products that Kilo Code currently integrates with:

  • 1
    Google Cloud Platform Reviews
    Top Pick

    Google Cloud Platform

    Google

    Free ($300 in free credits)
    60,421 Ratings
    See Software
    Learn More
    Google Cloud is an online service that lets you create everything from simple websites to complex apps for businesses of any size. Customers who are new to the system will receive $300 in credits for testing, deploying, and running workloads. Customers can use up to 25+ products free of charge. Use Google's core data analytics and machine learning. All enterprises can use it. It is secure and fully featured. Use big data to build better products and find answers faster. You can grow from prototypes to production and even to planet-scale without worrying about reliability, capacity or performance. Virtual machines with proven performance/price advantages, to a fully-managed app development platform. High performance, scalable, resilient object storage and databases. Google's private fibre network offers the latest software-defined networking solutions. Fully managed data warehousing and data exploration, Hadoop/Spark and messaging.
  • 2
    Visual Studio Code Reviews
    Top Pick
    Visual Studio Code is a highly extensible AI-powered code editor built for developers who demand flexibility and performance. It combines intelligent coding assistance, modern debugging tools, and collaboration features in one lightweight package. With Agent Mode, VS Code reads your codebase, runs terminal commands, and edits across files automatically until tasks are complete. Its Next Edit Suggestions feature predicts and completes your next move as you type, enhancing speed and code accuracy. The Model Context Protocol (MCP) enables developers to connect their favorite AI models—from OpenAI, Anthropic, Azure, or Google—and extend functionality through custom servers. Developers can work in any language, from JavaScript and Python to C#, Java, and Go, while leveraging over 75,000 extensions for added productivity. Seamless integration with GitHub Codespaces, cloud storage, and CI/CD tools allows teams to code, collaborate, and deploy anywhere. Open-source at its core, VS Code empowers both individuals and enterprises to innovate without limits.
  • 3
    Claude Sonnet 3.7 Reviews
    Claude Sonnet 3.7, a state-of-the-art AI model by Anthropic, is designed for versatility, offering users the option to switch between quick, efficient responses and deeper, more reflective answers. This dynamic model shines in complex problem-solving scenarios, where high-level reasoning and nuanced understanding are crucial. By allowing Claude to pause for self-reflection before answering, Sonnet 3.7 excels in tasks that demand deep analysis, such as coding, natural language processing, and critical thinking applications. Its flexibility makes it an invaluable tool for professionals and organizations looking for an adaptable AI that delivers both speed and thoughtful insights.
  • 4
    Grok Code Fast 1 Reviews

    Grok Code Fast 1

    xAI

    $0.20 per million input tokens
    Grok Code Fast 1 introduces a new class of coding-focused AI models that prioritize responsiveness, affordability, and real-world usability. Tailored for agentic coding platforms, it eliminates the lag developers often experience with reasoning loops and tool calls, creating a smoother workflow in IDEs. Its architecture was trained on a carefully curated mix of programming content and fine-tuned on real pull requests to reflect authentic development practices. With proficiency across multiple languages, including Python, Rust, TypeScript, C++, Java, and Go, it adapts to full-stack development scenarios. Grok Code Fast 1 excels in speed, processing nearly 190 tokens per second while maintaining reliable performance across bug fixes, code reviews, and project generation. Pricing makes it widely accessible at $0.20 per million input tokens, $1.50 per million output tokens, and just $0.02 for cached inputs. Early testers, including GitHub Copilot and Cursor users, praise its responsiveness and quality. For developers seeking a reliable coding assistant that’s both fast and cost-effective, Grok Code Fast 1 is a daily driver built for practical software engineering needs.
  • 5
    MiniMax M2 Reviews

    MiniMax M2

    MiniMax

    $0.30 per million input tokens
    MiniMax M2 is an open-source foundational model tailored for agent-driven applications and coding tasks, achieving an innovative equilibrium of efficiency, velocity, and affordability. It shines in comprehensive development environments, adeptly managing programming tasks, invoking tools, and executing intricate, multi-step processes, complete with features like Python integration, while offering impressive inference speeds of approximately 100 tokens per second and competitive API pricing at around 8% of similar proprietary models. The model includes a "Lightning Mode" designed for rapid, streamlined agent operations, alongside a "Pro Mode" aimed at thorough full-stack development, report creation, and the orchestration of web-based tools; its weights are entirely open source, allowing for local deployment via vLLM or SGLang. MiniMax M2 stands out as a model ready for production use, empowering agents to autonomously perform tasks such as data analysis, software development, tool orchestration, and implementing large-scale, multi-step logic across real organizational contexts. With its advanced capabilities, this model is poised to revolutionize the way developers approach complex programming challenges.
  • 6
    GLM-4.6 Reviews
    GLM-4.6 builds upon the foundations laid by its predecessor, showcasing enhanced reasoning, coding, and agent capabilities, resulting in notable advancements in inferential accuracy, improved tool usage during reasoning tasks, and a more seamless integration within agent frameworks. In comprehensive benchmark evaluations that assess reasoning, coding, and agent performance, GLM-4.6 surpasses GLM-4.5 and competes robustly against other models like DeepSeek-V3.2-Exp and Claude Sonnet 4, although it still lags behind Claude Sonnet 4.5 in terms of coding capabilities. Furthermore, when subjected to practical tests utilizing an extensive “CC-Bench” suite that includes tasks in front-end development, tool creation, data analysis, and algorithmic challenges, GLM-4.6 outperforms GLM-4.5 while nearing parity with Claude Sonnet 4, achieving victory in approximately 48.6% of direct comparisons and demonstrating around 15% improved token efficiency. This latest model is accessible through the Z.ai API, providing developers the flexibility to implement it as either an LLM backend or as the core of an agent within the platform's API ecosystem. In addition, its advancements could significantly enhance productivity in various application domains, making it an attractive option for developers looking to leverage cutting-edge AI technology.
  • Previous
  • You're on page 1
  • Next