Intel Open Edge PlatformIntel
|
||||||
Related Products
|
||||||
About
The Intel Open Edge Platform simplifies the development, deployment, and scaling of AI and edge computing solutions on standard hardware with cloud-like efficiency. It provides a curated set of components and workflows that accelerate AI model creation, optimization, and application development. From vision models to generative AI and large language models (LLM), the platform offers tools to streamline model training and inference. By integrating Intel’s OpenVINO toolkit, it ensures enhanced performance on Intel CPUs, GPUs, and VPUs, allowing organizations to bring AI applications to the edge with ease.
|
About
WebLLM is a high-performance, in-browser language model inference engine that leverages WebGPU for hardware acceleration, enabling powerful LLM operations directly within web browsers without server-side processing. It offers full OpenAI API compatibility, allowing seamless integration with functionalities such as JSON mode, function-calling, and streaming. WebLLM natively supports a range of models, including Llama, Phi, Gemma, RedPajama, Mistral, and Qwen, making it versatile for various AI tasks. Users can easily integrate and deploy custom models in MLC format, adapting WebLLM to specific needs and scenarios. The platform facilitates plug-and-play integration through package managers like NPM and Yarn, or directly via CDN, complemented by comprehensive examples and a modular design for connecting with UI components. It supports streaming chat completions for real-time output generation, enhancing interactive applications like chatbots and virtual assistants.
|
|||||
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
|||||
Audience
Businesses and developers looking for a powerful, scalable solution to build and deploy AI applications at the edge, leveraging Intel’s optimized hardware and cloud-like simplicity for edge computing
|
Audience
Developers seeking a tool to implement high-performance, in-browser language model inference without relying on server-side processing
|
|||||
Support
Phone Support
24/7 Live Support
Online
|
Support
Phone Support
24/7 Live Support
Online
|
|||||
API
Offers API
|
API
Offers API
|
|||||
Screenshots and Videos |
Screenshots and Videos |
|||||
Pricing
No information available.
Free Version
Free Trial
|
Pricing
Free
Free Version
Free Trial
|
|||||
Reviews/
|
Reviews/
|
|||||
Training
Documentation
Webinars
Live Online
In Person
|
Training
Documentation
Webinars
Live Online
In Person
|
|||||
Company InformationIntel
Founded: 1968
United States
www.intel.com/content/www/us/en/developer/tools/tiber/edge-platform/overview.html
|
Company InformationWebLLM
webllm.mlc.ai/
|
|||||
Alternatives |
Alternatives |
|||||
|
|
||||||
|
|
|
|||||
|
|
||||||
Categories |
Categories |
|||||
Integrations
Codestral
Dolly
Gemma
Google Cloud Confidential VMs
JupyterLab
Le Chat
Llama
Llama 2
Llama 3
Llama 3.1
|
Integrations
Codestral
Dolly
Gemma
Google Cloud Confidential VMs
JupyterLab
Le Chat
Llama
Llama 2
Llama 3
Llama 3.1
|
|||||
|
|
|