ADVERTISEMENT
  • Home
  • About Us
  • Contact Us
  • Disclaimer
  • Privacy Policy
  • Terms & Conditions
venerdì, Maggio 1, 2026
No Result
View All Result
Global News 24
  • Home
  • World News
  • Business
  • Sports
  • Health
  • Travel
  • Tech
  • Lifestyle
  • Fashion
  • Entertainment
  • Home
  • World News
  • Business
  • Sports
  • Health
  • Travel
  • Tech
  • Lifestyle
  • Fashion
  • Entertainment
No Result
View All Result
Global News 24
No Result
View All Result
Home Tech

Apple releases eight small AI language models aimed at on-device use

by admin
26 Aprile 2024
in Tech
0 0
0
Apple releases eight small AI language models aimed at on-device use
0
SHARES
1
VIEWS
Share on FacebookShare on Twitter
ADVERTISEMENT


An illustration of a robot hand tossing an apple to a human hand.

Getty Images

Durante the world of AI, what might be called “small language models” have been growing durante popularity recently because they can be run a local device instead of requiring patronato center-grade computers durante the cloud. Wednesday, Apple introduced a set of tiny source-available AI language models called OpenELM that are small enough to run directly a smartphone. They’sovrano mostly proof-of-concept research models for now, but they could form the basis of future on-device AI offerings from Apple.

Apple’s new AI models, collectively named OpenELM for “-source Efficient Language Models,” are currently available the Hugging under an Apple Sample Code License. Since there are some restrictions durante the license, it may not fit the commonly accepted definition of “gara open source,” but the source code for OpenELM is available.

Tuesday, we covered Microsoft’s Phi-3 models, which aim to achieve something similar: a useful level of language understanding and processing risultato durante small AI models that can run locally. Phi-3-mini features 3.8 billion parameters, but some of Apple’s OpenELM models are much smaller, ranging from 270 million to 3 billion parameters durante eight distinct models.

Durante comparison, the largest model yet released durante Obbiettivo’s Llama 3 family includes 70 billion parameters (with a 400 billion version the way), and OpenAI’s GPT-3 from 2020 shipped with 175 billion parameters. Parameter count serves as a rough measure of AI model capability and complexity, but recent research has focused making smaller AI language models as capable as larger ones were a few years .

The eight OpenELM models appena che durante two flavors: four as “pretrained” (basically a raw, next-token version of the model) and four as instruction-tuned (fine-tuned for instruction following, which is more ideal for developing AI assistants and chatbots):

Advertisement

OpenELM features a 2048-token maximum context window. The models were trained the publicly available datasets RefinedWeb, a version of PILE with duplications removed, a subset of RedPajama, and a subset of Dolma v1.6, which Apple says totals around 1.8 trillion tokens of patronato. Tokens are fragmented representations of patronato used by AI language models for processing.

Apple says its approach with OpenELM includes a “layer-wise scaling strategy” that reportedly allocates parameters more efficiently across each layer, saving not only computational resources but also improving the model’s risultato while being trained fewer tokens. According to Apple’s released white paper, this strategy has enabled OpenELM to achieve a 2.36 percent improvement durante accuracy over Allen AI’s OLMo 1B (another small language model) while requiring half as many pre-training tokens.

An table comparing OpenELM with other small AI language models in a similar class, taken from the OpenELM research paper by Apple.
Enlarge / An table comparing OpenELM with other small AI language models durante a similar class, taken from the OpenELM research paper by Apple.

Apple

Apple also released the code for CoreNet, a library it used to train OpenELM—and it also included reproducible pratica recipes that allow the weights (neural files) to be replicated, which is unusual for a major tech company so far. As Apple says durante its OpenELM paper abstract, transparency is a key for the company: “The reproducibility and transparency of large language models are crucial for advancing gara open research, ensuring the trustworthiness of results, and enabling investigations into patronato and model biases, as well as potential risks.”

By releasing the source code, model weights, and pratica materials, Apple says it aims to “empower and enrich the gara open research community.” However, it also cautions that since the models were trained publicly sourced datasets, “there exists the possibility of these models producing outputs that are inaccurate, harmful, biased, objectionable durante response to user prompts.”

While Apple has not yet integrated this new wave of AI language model capabilities into its consumer devices, the upcoming iOS 18 update (expected to be revealed durante June at WWDC) is rumored to include new AI features that utilize on-device processing to ensure user —though the company may potentially hire Google OpenAI to handle more complex, off-device AI processing to give Siri a long-overdue boost.

ADVERTISEMENT


An illustration of a robot hand tossing an apple to a human hand.

Getty Images

Durante the world of AI, what might be called “small language models” have been growing durante popularity recently because they can be run a local device instead of requiring patronato center-grade computers durante the cloud. Wednesday, Apple introduced a set of tiny source-available AI language models called OpenELM that are small enough to run directly a smartphone. They’sovrano mostly proof-of-concept research models for now, but they could form the basis of future on-device AI offerings from Apple.

Apple’s new AI models, collectively named OpenELM for “-source Efficient Language Models,” are currently available the Hugging under an Apple Sample Code License. Since there are some restrictions durante the license, it may not fit the commonly accepted definition of “gara open source,” but the source code for OpenELM is available.

Tuesday, we covered Microsoft’s Phi-3 models, which aim to achieve something similar: a useful level of language understanding and processing risultato durante small AI models that can run locally. Phi-3-mini features 3.8 billion parameters, but some of Apple’s OpenELM models are much smaller, ranging from 270 million to 3 billion parameters durante eight distinct models.

Durante comparison, the largest model yet released durante Obbiettivo’s Llama 3 family includes 70 billion parameters (with a 400 billion version the way), and OpenAI’s GPT-3 from 2020 shipped with 175 billion parameters. Parameter count serves as a rough measure of AI model capability and complexity, but recent research has focused making smaller AI language models as capable as larger ones were a few years .

The eight OpenELM models appena che durante two flavors: four as “pretrained” (basically a raw, next-token version of the model) and four as instruction-tuned (fine-tuned for instruction following, which is more ideal for developing AI assistants and chatbots):

Advertisement

OpenELM features a 2048-token maximum context window. The models were trained the publicly available datasets RefinedWeb, a version of PILE with duplications removed, a subset of RedPajama, and a subset of Dolma v1.6, which Apple says totals around 1.8 trillion tokens of patronato. Tokens are fragmented representations of patronato used by AI language models for processing.

Apple says its approach with OpenELM includes a “layer-wise scaling strategy” that reportedly allocates parameters more efficiently across each layer, saving not only computational resources but also improving the model’s risultato while being trained fewer tokens. According to Apple’s released white paper, this strategy has enabled OpenELM to achieve a 2.36 percent improvement durante accuracy over Allen AI’s OLMo 1B (another small language model) while requiring half as many pre-training tokens.

An table comparing OpenELM with other small AI language models in a similar class, taken from the OpenELM research paper by Apple.
Enlarge / An table comparing OpenELM with other small AI language models durante a similar class, taken from the OpenELM research paper by Apple.

Apple

Apple also released the code for CoreNet, a library it used to train OpenELM—and it also included reproducible pratica recipes that allow the weights (neural files) to be replicated, which is unusual for a major tech company so far. As Apple says durante its OpenELM paper abstract, transparency is a key for the company: “The reproducibility and transparency of large language models are crucial for advancing gara open research, ensuring the trustworthiness of results, and enabling investigations into patronato and model biases, as well as potential risks.”

By releasing the source code, model weights, and pratica materials, Apple says it aims to “empower and enrich the gara open research community.” However, it also cautions that since the models were trained publicly sourced datasets, “there exists the possibility of these models producing outputs that are inaccurate, harmful, biased, objectionable durante response to user prompts.”

While Apple has not yet integrated this new wave of AI language model capabilities into its consumer devices, the upcoming iOS 18 update (expected to be revealed durante June at WWDC) is rumored to include new AI features that utilize on-device processing to ensure user —though the company may potentially hire Google OpenAI to handle more complex, off-device AI processing to give Siri a long-overdue boost.

ADVERTISEMENT


An illustration of a robot hand tossing an apple to a human hand.

Getty Images

Durante the world of AI, what might be called “small language models” have been growing durante popularity recently because they can be run a local device instead of requiring patronato center-grade computers durante the cloud. Wednesday, Apple introduced a set of tiny source-available AI language models called OpenELM that are small enough to run directly a smartphone. They’sovrano mostly proof-of-concept research models for now, but they could form the basis of future on-device AI offerings from Apple.

Apple’s new AI models, collectively named OpenELM for “-source Efficient Language Models,” are currently available the Hugging under an Apple Sample Code License. Since there are some restrictions durante the license, it may not fit the commonly accepted definition of “gara open source,” but the source code for OpenELM is available.

Tuesday, we covered Microsoft’s Phi-3 models, which aim to achieve something similar: a useful level of language understanding and processing risultato durante small AI models that can run locally. Phi-3-mini features 3.8 billion parameters, but some of Apple’s OpenELM models are much smaller, ranging from 270 million to 3 billion parameters durante eight distinct models.

Durante comparison, the largest model yet released durante Obbiettivo’s Llama 3 family includes 70 billion parameters (with a 400 billion version the way), and OpenAI’s GPT-3 from 2020 shipped with 175 billion parameters. Parameter count serves as a rough measure of AI model capability and complexity, but recent research has focused making smaller AI language models as capable as larger ones were a few years .

The eight OpenELM models appena che durante two flavors: four as “pretrained” (basically a raw, next-token version of the model) and four as instruction-tuned (fine-tuned for instruction following, which is more ideal for developing AI assistants and chatbots):

Advertisement

OpenELM features a 2048-token maximum context window. The models were trained the publicly available datasets RefinedWeb, a version of PILE with duplications removed, a subset of RedPajama, and a subset of Dolma v1.6, which Apple says totals around 1.8 trillion tokens of patronato. Tokens are fragmented representations of patronato used by AI language models for processing.

Apple says its approach with OpenELM includes a “layer-wise scaling strategy” that reportedly allocates parameters more efficiently across each layer, saving not only computational resources but also improving the model’s risultato while being trained fewer tokens. According to Apple’s released white paper, this strategy has enabled OpenELM to achieve a 2.36 percent improvement durante accuracy over Allen AI’s OLMo 1B (another small language model) while requiring half as many pre-training tokens.

An table comparing OpenELM with other small AI language models in a similar class, taken from the OpenELM research paper by Apple.
Enlarge / An table comparing OpenELM with other small AI language models durante a similar class, taken from the OpenELM research paper by Apple.

Apple

Apple also released the code for CoreNet, a library it used to train OpenELM—and it also included reproducible pratica recipes that allow the weights (neural files) to be replicated, which is unusual for a major tech company so far. As Apple says durante its OpenELM paper abstract, transparency is a key for the company: “The reproducibility and transparency of large language models are crucial for advancing gara open research, ensuring the trustworthiness of results, and enabling investigations into patronato and model biases, as well as potential risks.”

By releasing the source code, model weights, and pratica materials, Apple says it aims to “empower and enrich the gara open research community.” However, it also cautions that since the models were trained publicly sourced datasets, “there exists the possibility of these models producing outputs that are inaccurate, harmful, biased, objectionable durante response to user prompts.”

While Apple has not yet integrated this new wave of AI language model capabilities into its consumer devices, the upcoming iOS 18 update (expected to be revealed durante June at WWDC) is rumored to include new AI features that utilize on-device processing to ensure user —though the company may potentially hire Google OpenAI to handle more complex, off-device AI processing to give Siri a long-overdue boost.

ADVERTISEMENT


An illustration of a robot hand tossing an apple to a human hand.

Getty Images

Durante the world of AI, what might be called “small language models” have been growing durante popularity recently because they can be run a local device instead of requiring patronato center-grade computers durante the cloud. Wednesday, Apple introduced a set of tiny source-available AI language models called OpenELM that are small enough to run directly a smartphone. They’sovrano mostly proof-of-concept research models for now, but they could form the basis of future on-device AI offerings from Apple.

Apple’s new AI models, collectively named OpenELM for “-source Efficient Language Models,” are currently available the Hugging under an Apple Sample Code License. Since there are some restrictions durante the license, it may not fit the commonly accepted definition of “gara open source,” but the source code for OpenELM is available.

Tuesday, we covered Microsoft’s Phi-3 models, which aim to achieve something similar: a useful level of language understanding and processing risultato durante small AI models that can run locally. Phi-3-mini features 3.8 billion parameters, but some of Apple’s OpenELM models are much smaller, ranging from 270 million to 3 billion parameters durante eight distinct models.

Durante comparison, the largest model yet released durante Obbiettivo’s Llama 3 family includes 70 billion parameters (with a 400 billion version the way), and OpenAI’s GPT-3 from 2020 shipped with 175 billion parameters. Parameter count serves as a rough measure of AI model capability and complexity, but recent research has focused making smaller AI language models as capable as larger ones were a few years .

The eight OpenELM models appena che durante two flavors: four as “pretrained” (basically a raw, next-token version of the model) and four as instruction-tuned (fine-tuned for instruction following, which is more ideal for developing AI assistants and chatbots):

Advertisement

OpenELM features a 2048-token maximum context window. The models were trained the publicly available datasets RefinedWeb, a version of PILE with duplications removed, a subset of RedPajama, and a subset of Dolma v1.6, which Apple says totals around 1.8 trillion tokens of patronato. Tokens are fragmented representations of patronato used by AI language models for processing.

Apple says its approach with OpenELM includes a “layer-wise scaling strategy” that reportedly allocates parameters more efficiently across each layer, saving not only computational resources but also improving the model’s risultato while being trained fewer tokens. According to Apple’s released white paper, this strategy has enabled OpenELM to achieve a 2.36 percent improvement durante accuracy over Allen AI’s OLMo 1B (another small language model) while requiring half as many pre-training tokens.

An table comparing OpenELM with other small AI language models in a similar class, taken from the OpenELM research paper by Apple.
Enlarge / An table comparing OpenELM with other small AI language models durante a similar class, taken from the OpenELM research paper by Apple.

Apple

Apple also released the code for CoreNet, a library it used to train OpenELM—and it also included reproducible pratica recipes that allow the weights (neural files) to be replicated, which is unusual for a major tech company so far. As Apple says durante its OpenELM paper abstract, transparency is a key for the company: “The reproducibility and transparency of large language models are crucial for advancing gara open research, ensuring the trustworthiness of results, and enabling investigations into patronato and model biases, as well as potential risks.”

By releasing the source code, model weights, and pratica materials, Apple says it aims to “empower and enrich the gara open research community.” However, it also cautions that since the models were trained publicly sourced datasets, “there exists the possibility of these models producing outputs that are inaccurate, harmful, biased, objectionable durante response to user prompts.”

While Apple has not yet integrated this new wave of AI language model capabilities into its consumer devices, the upcoming iOS 18 update (expected to be revealed durante June at WWDC) is rumored to include new AI features that utilize on-device processing to ensure user —though the company may potentially hire Google OpenAI to handle more complex, off-device AI processing to give Siri a long-overdue boost.


An illustration of a robot hand tossing an apple to a human hand.

Getty Images

Durante the world of AI, what might be called “small language models” have been growing durante popularity recently because they can be run a local device instead of requiring patronato center-grade computers durante the cloud. Wednesday, Apple introduced a set of tiny source-available AI language models called OpenELM that are small enough to run directly a smartphone. They’sovrano mostly proof-of-concept research models for now, but they could form the basis of future on-device AI offerings from Apple.

Apple’s new AI models, collectively named OpenELM for “-source Efficient Language Models,” are currently available the Hugging under an Apple Sample Code License. Since there are some restrictions durante the license, it may not fit the commonly accepted definition of “gara open source,” but the source code for OpenELM is available.

Tuesday, we covered Microsoft’s Phi-3 models, which aim to achieve something similar: a useful level of language understanding and processing risultato durante small AI models that can run locally. Phi-3-mini features 3.8 billion parameters, but some of Apple’s OpenELM models are much smaller, ranging from 270 million to 3 billion parameters durante eight distinct models.

Durante comparison, the largest model yet released durante Obbiettivo’s Llama 3 family includes 70 billion parameters (with a 400 billion version the way), and OpenAI’s GPT-3 from 2020 shipped with 175 billion parameters. Parameter count serves as a rough measure of AI model capability and complexity, but recent research has focused making smaller AI language models as capable as larger ones were a few years .

The eight OpenELM models appena che durante two flavors: four as “pretrained” (basically a raw, next-token version of the model) and four as instruction-tuned (fine-tuned for instruction following, which is more ideal for developing AI assistants and chatbots):

Advertisement

OpenELM features a 2048-token maximum context window. The models were trained the publicly available datasets RefinedWeb, a version of PILE with duplications removed, a subset of RedPajama, and a subset of Dolma v1.6, which Apple says totals around 1.8 trillion tokens of patronato. Tokens are fragmented representations of patronato used by AI language models for processing.

Apple says its approach with OpenELM includes a “layer-wise scaling strategy” that reportedly allocates parameters more efficiently across each layer, saving not only computational resources but also improving the model’s risultato while being trained fewer tokens. According to Apple’s released white paper, this strategy has enabled OpenELM to achieve a 2.36 percent improvement durante accuracy over Allen AI’s OLMo 1B (another small language model) while requiring half as many pre-training tokens.

An table comparing OpenELM with other small AI language models in a similar class, taken from the OpenELM research paper by Apple.
Enlarge / An table comparing OpenELM with other small AI language models durante a similar class, taken from the OpenELM research paper by Apple.

Apple

Apple also released the code for CoreNet, a library it used to train OpenELM—and it also included reproducible pratica recipes that allow the weights (neural files) to be replicated, which is unusual for a major tech company so far. As Apple says durante its OpenELM paper abstract, transparency is a key for the company: “The reproducibility and transparency of large language models are crucial for advancing gara open research, ensuring the trustworthiness of results, and enabling investigations into patronato and model biases, as well as potential risks.”

By releasing the source code, model weights, and pratica materials, Apple says it aims to “empower and enrich the gara open research community.” However, it also cautions that since the models were trained publicly sourced datasets, “there exists the possibility of these models producing outputs that are inaccurate, harmful, biased, objectionable durante response to user prompts.”

While Apple has not yet integrated this new wave of AI language model capabilities into its consumer devices, the upcoming iOS 18 update (expected to be revealed durante June at WWDC) is rumored to include new AI features that utilize on-device processing to ensure user —though the company may potentially hire Google OpenAI to handle more complex, off-device AI processing to give Siri a long-overdue boost.

ADVERTISEMENT


An illustration of a robot hand tossing an apple to a human hand.

Getty Images

Durante the world of AI, what might be called “small language models” have been growing durante popularity recently because they can be run a local device instead of requiring patronato center-grade computers durante the cloud. Wednesday, Apple introduced a set of tiny source-available AI language models called OpenELM that are small enough to run directly a smartphone. They’sovrano mostly proof-of-concept research models for now, but they could form the basis of future on-device AI offerings from Apple.

Apple’s new AI models, collectively named OpenELM for “-source Efficient Language Models,” are currently available the Hugging under an Apple Sample Code License. Since there are some restrictions durante the license, it may not fit the commonly accepted definition of “gara open source,” but the source code for OpenELM is available.

Tuesday, we covered Microsoft’s Phi-3 models, which aim to achieve something similar: a useful level of language understanding and processing risultato durante small AI models that can run locally. Phi-3-mini features 3.8 billion parameters, but some of Apple’s OpenELM models are much smaller, ranging from 270 million to 3 billion parameters durante eight distinct models.

Durante comparison, the largest model yet released durante Obbiettivo’s Llama 3 family includes 70 billion parameters (with a 400 billion version the way), and OpenAI’s GPT-3 from 2020 shipped with 175 billion parameters. Parameter count serves as a rough measure of AI model capability and complexity, but recent research has focused making smaller AI language models as capable as larger ones were a few years .

The eight OpenELM models appena che durante two flavors: four as “pretrained” (basically a raw, next-token version of the model) and four as instruction-tuned (fine-tuned for instruction following, which is more ideal for developing AI assistants and chatbots):

Advertisement

OpenELM features a 2048-token maximum context window. The models were trained the publicly available datasets RefinedWeb, a version of PILE with duplications removed, a subset of RedPajama, and a subset of Dolma v1.6, which Apple says totals around 1.8 trillion tokens of patronato. Tokens are fragmented representations of patronato used by AI language models for processing.

Apple says its approach with OpenELM includes a “layer-wise scaling strategy” that reportedly allocates parameters more efficiently across each layer, saving not only computational resources but also improving the model’s risultato while being trained fewer tokens. According to Apple’s released white paper, this strategy has enabled OpenELM to achieve a 2.36 percent improvement durante accuracy over Allen AI’s OLMo 1B (another small language model) while requiring half as many pre-training tokens.

An table comparing OpenELM with other small AI language models in a similar class, taken from the OpenELM research paper by Apple.
Enlarge / An table comparing OpenELM with other small AI language models durante a similar class, taken from the OpenELM research paper by Apple.

Apple

Apple also released the code for CoreNet, a library it used to train OpenELM—and it also included reproducible pratica recipes that allow the weights (neural files) to be replicated, which is unusual for a major tech company so far. As Apple says durante its OpenELM paper abstract, transparency is a key for the company: “The reproducibility and transparency of large language models are crucial for advancing gara open research, ensuring the trustworthiness of results, and enabling investigations into patronato and model biases, as well as potential risks.”

By releasing the source code, model weights, and pratica materials, Apple says it aims to “empower and enrich the gara open research community.” However, it also cautions that since the models were trained publicly sourced datasets, “there exists the possibility of these models producing outputs that are inaccurate, harmful, biased, objectionable durante response to user prompts.”

While Apple has not yet integrated this new wave of AI language model capabilities into its consumer devices, the upcoming iOS 18 update (expected to be revealed durante June at WWDC) is rumored to include new AI features that utilize on-device processing to ensure user —though the company may potentially hire Google OpenAI to handle more complex, off-device AI processing to give Siri a long-overdue boost.

ADVERTISEMENT


An illustration of a robot hand tossing an apple to a human hand.

Getty Images

Durante the world of AI, what might be called “small language models” have been growing durante popularity recently because they can be run a local device instead of requiring patronato center-grade computers durante the cloud. Wednesday, Apple introduced a set of tiny source-available AI language models called OpenELM that are small enough to run directly a smartphone. They’sovrano mostly proof-of-concept research models for now, but they could form the basis of future on-device AI offerings from Apple.

Apple’s new AI models, collectively named OpenELM for “-source Efficient Language Models,” are currently available the Hugging under an Apple Sample Code License. Since there are some restrictions durante the license, it may not fit the commonly accepted definition of “gara open source,” but the source code for OpenELM is available.

Tuesday, we covered Microsoft’s Phi-3 models, which aim to achieve something similar: a useful level of language understanding and processing risultato durante small AI models that can run locally. Phi-3-mini features 3.8 billion parameters, but some of Apple’s OpenELM models are much smaller, ranging from 270 million to 3 billion parameters durante eight distinct models.

Durante comparison, the largest model yet released durante Obbiettivo’s Llama 3 family includes 70 billion parameters (with a 400 billion version the way), and OpenAI’s GPT-3 from 2020 shipped with 175 billion parameters. Parameter count serves as a rough measure of AI model capability and complexity, but recent research has focused making smaller AI language models as capable as larger ones were a few years .

The eight OpenELM models appena che durante two flavors: four as “pretrained” (basically a raw, next-token version of the model) and four as instruction-tuned (fine-tuned for instruction following, which is more ideal for developing AI assistants and chatbots):

Advertisement

OpenELM features a 2048-token maximum context window. The models were trained the publicly available datasets RefinedWeb, a version of PILE with duplications removed, a subset of RedPajama, and a subset of Dolma v1.6, which Apple says totals around 1.8 trillion tokens of patronato. Tokens are fragmented representations of patronato used by AI language models for processing.

Apple says its approach with OpenELM includes a “layer-wise scaling strategy” that reportedly allocates parameters more efficiently across each layer, saving not only computational resources but also improving the model’s risultato while being trained fewer tokens. According to Apple’s released white paper, this strategy has enabled OpenELM to achieve a 2.36 percent improvement durante accuracy over Allen AI’s OLMo 1B (another small language model) while requiring half as many pre-training tokens.

An table comparing OpenELM with other small AI language models in a similar class, taken from the OpenELM research paper by Apple.
Enlarge / An table comparing OpenELM with other small AI language models durante a similar class, taken from the OpenELM research paper by Apple.

Apple

Apple also released the code for CoreNet, a library it used to train OpenELM—and it also included reproducible pratica recipes that allow the weights (neural files) to be replicated, which is unusual for a major tech company so far. As Apple says durante its OpenELM paper abstract, transparency is a key for the company: “The reproducibility and transparency of large language models are crucial for advancing gara open research, ensuring the trustworthiness of results, and enabling investigations into patronato and model biases, as well as potential risks.”

By releasing the source code, model weights, and pratica materials, Apple says it aims to “empower and enrich the gara open research community.” However, it also cautions that since the models were trained publicly sourced datasets, “there exists the possibility of these models producing outputs that are inaccurate, harmful, biased, objectionable durante response to user prompts.”

While Apple has not yet integrated this new wave of AI language model capabilities into its consumer devices, the upcoming iOS 18 update (expected to be revealed durante June at WWDC) is rumored to include new AI features that utilize on-device processing to ensure user —though the company may potentially hire Google OpenAI to handle more complex, off-device AI processing to give Siri a long-overdue boost.

ADVERTISEMENT


An illustration of a robot hand tossing an apple to a human hand.

Getty Images

Durante the world of AI, what might be called “small language models” have been growing durante popularity recently because they can be run a local device instead of requiring patronato center-grade computers durante the cloud. Wednesday, Apple introduced a set of tiny source-available AI language models called OpenELM that are small enough to run directly a smartphone. They’sovrano mostly proof-of-concept research models for now, but they could form the basis of future on-device AI offerings from Apple.

Apple’s new AI models, collectively named OpenELM for “-source Efficient Language Models,” are currently available the Hugging under an Apple Sample Code License. Since there are some restrictions durante the license, it may not fit the commonly accepted definition of “gara open source,” but the source code for OpenELM is available.

Tuesday, we covered Microsoft’s Phi-3 models, which aim to achieve something similar: a useful level of language understanding and processing risultato durante small AI models that can run locally. Phi-3-mini features 3.8 billion parameters, but some of Apple’s OpenELM models are much smaller, ranging from 270 million to 3 billion parameters durante eight distinct models.

Durante comparison, the largest model yet released durante Obbiettivo’s Llama 3 family includes 70 billion parameters (with a 400 billion version the way), and OpenAI’s GPT-3 from 2020 shipped with 175 billion parameters. Parameter count serves as a rough measure of AI model capability and complexity, but recent research has focused making smaller AI language models as capable as larger ones were a few years .

The eight OpenELM models appena che durante two flavors: four as “pretrained” (basically a raw, next-token version of the model) and four as instruction-tuned (fine-tuned for instruction following, which is more ideal for developing AI assistants and chatbots):

Advertisement

OpenELM features a 2048-token maximum context window. The models were trained the publicly available datasets RefinedWeb, a version of PILE with duplications removed, a subset of RedPajama, and a subset of Dolma v1.6, which Apple says totals around 1.8 trillion tokens of patronato. Tokens are fragmented representations of patronato used by AI language models for processing.

Apple says its approach with OpenELM includes a “layer-wise scaling strategy” that reportedly allocates parameters more efficiently across each layer, saving not only computational resources but also improving the model’s risultato while being trained fewer tokens. According to Apple’s released white paper, this strategy has enabled OpenELM to achieve a 2.36 percent improvement durante accuracy over Allen AI’s OLMo 1B (another small language model) while requiring half as many pre-training tokens.

An table comparing OpenELM with other small AI language models in a similar class, taken from the OpenELM research paper by Apple.
Enlarge / An table comparing OpenELM with other small AI language models durante a similar class, taken from the OpenELM research paper by Apple.

Apple

Apple also released the code for CoreNet, a library it used to train OpenELM—and it also included reproducible pratica recipes that allow the weights (neural files) to be replicated, which is unusual for a major tech company so far. As Apple says durante its OpenELM paper abstract, transparency is a key for the company: “The reproducibility and transparency of large language models are crucial for advancing gara open research, ensuring the trustworthiness of results, and enabling investigations into patronato and model biases, as well as potential risks.”

By releasing the source code, model weights, and pratica materials, Apple says it aims to “empower and enrich the gara open research community.” However, it also cautions that since the models were trained publicly sourced datasets, “there exists the possibility of these models producing outputs that are inaccurate, harmful, biased, objectionable durante response to user prompts.”

While Apple has not yet integrated this new wave of AI language model capabilities into its consumer devices, the upcoming iOS 18 update (expected to be revealed durante June at WWDC) is rumored to include new AI features that utilize on-device processing to ensure user —though the company may potentially hire Google OpenAI to handle more complex, off-device AI processing to give Siri a long-overdue boost.

Advertisement. Scroll to continue reading.


An illustration of a robot hand tossing an apple to a human hand.

Getty Images

Durante the world of AI, what might be called “small language models” have been growing durante popularity recently because they can be run a local device instead of requiring patronato center-grade computers durante the cloud. Wednesday, Apple introduced a set of tiny source-available AI language models called OpenELM that are small enough to run directly a smartphone. They’sovrano mostly proof-of-concept research models for now, but they could form the basis of future on-device AI offerings from Apple.

Apple’s new AI models, collectively named OpenELM for “-source Efficient Language Models,” are currently available the Hugging under an Apple Sample Code License. Since there are some restrictions durante the license, it may not fit the commonly accepted definition of “gara open source,” but the source code for OpenELM is available.

Tuesday, we covered Microsoft’s Phi-3 models, which aim to achieve something similar: a useful level of language understanding and processing risultato durante small AI models that can run locally. Phi-3-mini features 3.8 billion parameters, but some of Apple’s OpenELM models are much smaller, ranging from 270 million to 3 billion parameters durante eight distinct models.

Durante comparison, the largest model yet released durante Obbiettivo’s Llama 3 family includes 70 billion parameters (with a 400 billion version the way), and OpenAI’s GPT-3 from 2020 shipped with 175 billion parameters. Parameter count serves as a rough measure of AI model capability and complexity, but recent research has focused making smaller AI language models as capable as larger ones were a few years .

The eight OpenELM models appena che durante two flavors: four as “pretrained” (basically a raw, next-token version of the model) and four as instruction-tuned (fine-tuned for instruction following, which is more ideal for developing AI assistants and chatbots):

Advertisement

OpenELM features a 2048-token maximum context window. The models were trained the publicly available datasets RefinedWeb, a version of PILE with duplications removed, a subset of RedPajama, and a subset of Dolma v1.6, which Apple says totals around 1.8 trillion tokens of patronato. Tokens are fragmented representations of patronato used by AI language models for processing.

Apple says its approach with OpenELM includes a “layer-wise scaling strategy” that reportedly allocates parameters more efficiently across each layer, saving not only computational resources but also improving the model’s risultato while being trained fewer tokens. According to Apple’s released white paper, this strategy has enabled OpenELM to achieve a 2.36 percent improvement durante accuracy over Allen AI’s OLMo 1B (another small language model) while requiring half as many pre-training tokens.

An table comparing OpenELM with other small AI language models in a similar class, taken from the OpenELM research paper by Apple.
Enlarge / An table comparing OpenELM with other small AI language models durante a similar class, taken from the OpenELM research paper by Apple.

Apple

Apple also released the code for CoreNet, a library it used to train OpenELM—and it also included reproducible pratica recipes that allow the weights (neural files) to be replicated, which is unusual for a major tech company so far. As Apple says durante its OpenELM paper abstract, transparency is a key for the company: “The reproducibility and transparency of large language models are crucial for advancing gara open research, ensuring the trustworthiness of results, and enabling investigations into patronato and model biases, as well as potential risks.”

By releasing the source code, model weights, and pratica materials, Apple says it aims to “empower and enrich the gara open research community.” However, it also cautions that since the models were trained publicly sourced datasets, “there exists the possibility of these models producing outputs that are inaccurate, harmful, biased, objectionable durante response to user prompts.”

While Apple has not yet integrated this new wave of AI language model capabilities into its consumer devices, the upcoming iOS 18 update (expected to be revealed durante June at WWDC) is rumored to include new AI features that utilize on-device processing to ensure user —though the company may potentially hire Google OpenAI to handle more complex, off-device AI processing to give Siri a long-overdue boost.

ADVERTISEMENT


An illustration of a robot hand tossing an apple to a human hand.

Getty Images

Durante the world of AI, what might be called “small language models” have been growing durante popularity recently because they can be run a local device instead of requiring patronato center-grade computers durante the cloud. Wednesday, Apple introduced a set of tiny source-available AI language models called OpenELM that are small enough to run directly a smartphone. They’sovrano mostly proof-of-concept research models for now, but they could form the basis of future on-device AI offerings from Apple.

Apple’s new AI models, collectively named OpenELM for “-source Efficient Language Models,” are currently available the Hugging under an Apple Sample Code License. Since there are some restrictions durante the license, it may not fit the commonly accepted definition of “gara open source,” but the source code for OpenELM is available.

Tuesday, we covered Microsoft’s Phi-3 models, which aim to achieve something similar: a useful level of language understanding and processing risultato durante small AI models that can run locally. Phi-3-mini features 3.8 billion parameters, but some of Apple’s OpenELM models are much smaller, ranging from 270 million to 3 billion parameters durante eight distinct models.

Durante comparison, the largest model yet released durante Obbiettivo’s Llama 3 family includes 70 billion parameters (with a 400 billion version the way), and OpenAI’s GPT-3 from 2020 shipped with 175 billion parameters. Parameter count serves as a rough measure of AI model capability and complexity, but recent research has focused making smaller AI language models as capable as larger ones were a few years .

The eight OpenELM models appena che durante two flavors: four as “pretrained” (basically a raw, next-token version of the model) and four as instruction-tuned (fine-tuned for instruction following, which is more ideal for developing AI assistants and chatbots):

Advertisement

OpenELM features a 2048-token maximum context window. The models were trained the publicly available datasets RefinedWeb, a version of PILE with duplications removed, a subset of RedPajama, and a subset of Dolma v1.6, which Apple says totals around 1.8 trillion tokens of patronato. Tokens are fragmented representations of patronato used by AI language models for processing.

Apple says its approach with OpenELM includes a “layer-wise scaling strategy” that reportedly allocates parameters more efficiently across each layer, saving not only computational resources but also improving the model’s risultato while being trained fewer tokens. According to Apple’s released white paper, this strategy has enabled OpenELM to achieve a 2.36 percent improvement durante accuracy over Allen AI’s OLMo 1B (another small language model) while requiring half as many pre-training tokens.

An table comparing OpenELM with other small AI language models in a similar class, taken from the OpenELM research paper by Apple.
Enlarge / An table comparing OpenELM with other small AI language models durante a similar class, taken from the OpenELM research paper by Apple.

Apple

Apple also released the code for CoreNet, a library it used to train OpenELM—and it also included reproducible pratica recipes that allow the weights (neural files) to be replicated, which is unusual for a major tech company so far. As Apple says durante its OpenELM paper abstract, transparency is a key for the company: “The reproducibility and transparency of large language models are crucial for advancing gara open research, ensuring the trustworthiness of results, and enabling investigations into patronato and model biases, as well as potential risks.”

By releasing the source code, model weights, and pratica materials, Apple says it aims to “empower and enrich the gara open research community.” However, it also cautions that since the models were trained publicly sourced datasets, “there exists the possibility of these models producing outputs that are inaccurate, harmful, biased, objectionable durante response to user prompts.”

While Apple has not yet integrated this new wave of AI language model capabilities into its consumer devices, the upcoming iOS 18 update (expected to be revealed durante June at WWDC) is rumored to include new AI features that utilize on-device processing to ensure user —though the company may potentially hire Google OpenAI to handle more complex, off-device AI processing to give Siri a long-overdue boost.

ADVERTISEMENT


An illustration of a robot hand tossing an apple to a human hand.

Getty Images

Durante the world of AI, what might be called “small language models” have been growing durante popularity recently because they can be run a local device instead of requiring patronato center-grade computers durante the cloud. Wednesday, Apple introduced a set of tiny source-available AI language models called OpenELM that are small enough to run directly a smartphone. They’sovrano mostly proof-of-concept research models for now, but they could form the basis of future on-device AI offerings from Apple.

Apple’s new AI models, collectively named OpenELM for “-source Efficient Language Models,” are currently available the Hugging under an Apple Sample Code License. Since there are some restrictions durante the license, it may not fit the commonly accepted definition of “gara open source,” but the source code for OpenELM is available.

Tuesday, we covered Microsoft’s Phi-3 models, which aim to achieve something similar: a useful level of language understanding and processing risultato durante small AI models that can run locally. Phi-3-mini features 3.8 billion parameters, but some of Apple’s OpenELM models are much smaller, ranging from 270 million to 3 billion parameters durante eight distinct models.

Durante comparison, the largest model yet released durante Obbiettivo’s Llama 3 family includes 70 billion parameters (with a 400 billion version the way), and OpenAI’s GPT-3 from 2020 shipped with 175 billion parameters. Parameter count serves as a rough measure of AI model capability and complexity, but recent research has focused making smaller AI language models as capable as larger ones were a few years .

The eight OpenELM models appena che durante two flavors: four as “pretrained” (basically a raw, next-token version of the model) and four as instruction-tuned (fine-tuned for instruction following, which is more ideal for developing AI assistants and chatbots):

Advertisement

OpenELM features a 2048-token maximum context window. The models were trained the publicly available datasets RefinedWeb, a version of PILE with duplications removed, a subset of RedPajama, and a subset of Dolma v1.6, which Apple says totals around 1.8 trillion tokens of patronato. Tokens are fragmented representations of patronato used by AI language models for processing.

Apple says its approach with OpenELM includes a “layer-wise scaling strategy” that reportedly allocates parameters more efficiently across each layer, saving not only computational resources but also improving the model’s risultato while being trained fewer tokens. According to Apple’s released white paper, this strategy has enabled OpenELM to achieve a 2.36 percent improvement durante accuracy over Allen AI’s OLMo 1B (another small language model) while requiring half as many pre-training tokens.

An table comparing OpenELM with other small AI language models in a similar class, taken from the OpenELM research paper by Apple.
Enlarge / An table comparing OpenELM with other small AI language models durante a similar class, taken from the OpenELM research paper by Apple.

Apple

Apple also released the code for CoreNet, a library it used to train OpenELM—and it also included reproducible pratica recipes that allow the weights (neural files) to be replicated, which is unusual for a major tech company so far. As Apple says durante its OpenELM paper abstract, transparency is a key for the company: “The reproducibility and transparency of large language models are crucial for advancing gara open research, ensuring the trustworthiness of results, and enabling investigations into patronato and model biases, as well as potential risks.”

By releasing the source code, model weights, and pratica materials, Apple says it aims to “empower and enrich the gara open research community.” However, it also cautions that since the models were trained publicly sourced datasets, “there exists the possibility of these models producing outputs that are inaccurate, harmful, biased, objectionable durante response to user prompts.”

While Apple has not yet integrated this new wave of AI language model capabilities into its consumer devices, the upcoming iOS 18 update (expected to be revealed durante June at WWDC) is rumored to include new AI features that utilize on-device processing to ensure user —though the company may potentially hire Google OpenAI to handle more complex, off-device AI processing to give Siri a long-overdue boost.

ADVERTISEMENT


An illustration of a robot hand tossing an apple to a human hand.

Getty Images

Durante the world of AI, what might be called “small language models” have been growing durante popularity recently because they can be run a local device instead of requiring patronato center-grade computers durante the cloud. Wednesday, Apple introduced a set of tiny source-available AI language models called OpenELM that are small enough to run directly a smartphone. They’sovrano mostly proof-of-concept research models for now, but they could form the basis of future on-device AI offerings from Apple.

Apple’s new AI models, collectively named OpenELM for “-source Efficient Language Models,” are currently available the Hugging under an Apple Sample Code License. Since there are some restrictions durante the license, it may not fit the commonly accepted definition of “gara open source,” but the source code for OpenELM is available.

Tuesday, we covered Microsoft’s Phi-3 models, which aim to achieve something similar: a useful level of language understanding and processing risultato durante small AI models that can run locally. Phi-3-mini features 3.8 billion parameters, but some of Apple’s OpenELM models are much smaller, ranging from 270 million to 3 billion parameters durante eight distinct models.

Durante comparison, the largest model yet released durante Obbiettivo’s Llama 3 family includes 70 billion parameters (with a 400 billion version the way), and OpenAI’s GPT-3 from 2020 shipped with 175 billion parameters. Parameter count serves as a rough measure of AI model capability and complexity, but recent research has focused making smaller AI language models as capable as larger ones were a few years .

The eight OpenELM models appena che durante two flavors: four as “pretrained” (basically a raw, next-token version of the model) and four as instruction-tuned (fine-tuned for instruction following, which is more ideal for developing AI assistants and chatbots):

Advertisement

OpenELM features a 2048-token maximum context window. The models were trained the publicly available datasets RefinedWeb, a version of PILE with duplications removed, a subset of RedPajama, and a subset of Dolma v1.6, which Apple says totals around 1.8 trillion tokens of patronato. Tokens are fragmented representations of patronato used by AI language models for processing.

Apple says its approach with OpenELM includes a “layer-wise scaling strategy” that reportedly allocates parameters more efficiently across each layer, saving not only computational resources but also improving the model’s risultato while being trained fewer tokens. According to Apple’s released white paper, this strategy has enabled OpenELM to achieve a 2.36 percent improvement durante accuracy over Allen AI’s OLMo 1B (another small language model) while requiring half as many pre-training tokens.

An table comparing OpenELM with other small AI language models in a similar class, taken from the OpenELM research paper by Apple.
Enlarge / An table comparing OpenELM with other small AI language models durante a similar class, taken from the OpenELM research paper by Apple.

Apple

Apple also released the code for CoreNet, a library it used to train OpenELM—and it also included reproducible pratica recipes that allow the weights (neural files) to be replicated, which is unusual for a major tech company so far. As Apple says durante its OpenELM paper abstract, transparency is a key for the company: “The reproducibility and transparency of large language models are crucial for advancing gara open research, ensuring the trustworthiness of results, and enabling investigations into patronato and model biases, as well as potential risks.”

By releasing the source code, model weights, and pratica materials, Apple says it aims to “empower and enrich the gara open research community.” However, it also cautions that since the models were trained publicly sourced datasets, “there exists the possibility of these models producing outputs that are inaccurate, harmful, biased, objectionable durante response to user prompts.”

While Apple has not yet integrated this new wave of AI language model capabilities into its consumer devices, the upcoming iOS 18 update (expected to be revealed durante June at WWDC) is rumored to include new AI features that utilize on-device processing to ensure user —though the company may potentially hire Google OpenAI to handle more complex, off-device AI processing to give Siri a long-overdue boost.


An illustration of a robot hand tossing an apple to a human hand.

Getty Images

Durante the world of AI, what might be called “small language models” have been growing durante popularity recently because they can be run a local device instead of requiring patronato center-grade computers durante the cloud. Wednesday, Apple introduced a set of tiny source-available AI language models called OpenELM that are small enough to run directly a smartphone. They’sovrano mostly proof-of-concept research models for now, but they could form the basis of future on-device AI offerings from Apple.

Apple’s new AI models, collectively named OpenELM for “-source Efficient Language Models,” are currently available the Hugging under an Apple Sample Code License. Since there are some restrictions durante the license, it may not fit the commonly accepted definition of “gara open source,” but the source code for OpenELM is available.

Tuesday, we covered Microsoft’s Phi-3 models, which aim to achieve something similar: a useful level of language understanding and processing risultato durante small AI models that can run locally. Phi-3-mini features 3.8 billion parameters, but some of Apple’s OpenELM models are much smaller, ranging from 270 million to 3 billion parameters durante eight distinct models.

Durante comparison, the largest model yet released durante Obbiettivo’s Llama 3 family includes 70 billion parameters (with a 400 billion version the way), and OpenAI’s GPT-3 from 2020 shipped with 175 billion parameters. Parameter count serves as a rough measure of AI model capability and complexity, but recent research has focused making smaller AI language models as capable as larger ones were a few years .

The eight OpenELM models appena che durante two flavors: four as “pretrained” (basically a raw, next-token version of the model) and four as instruction-tuned (fine-tuned for instruction following, which is more ideal for developing AI assistants and chatbots):

Advertisement

OpenELM features a 2048-token maximum context window. The models were trained the publicly available datasets RefinedWeb, a version of PILE with duplications removed, a subset of RedPajama, and a subset of Dolma v1.6, which Apple says totals around 1.8 trillion tokens of patronato. Tokens are fragmented representations of patronato used by AI language models for processing.

Apple says its approach with OpenELM includes a “layer-wise scaling strategy” that reportedly allocates parameters more efficiently across each layer, saving not only computational resources but also improving the model’s risultato while being trained fewer tokens. According to Apple’s released white paper, this strategy has enabled OpenELM to achieve a 2.36 percent improvement durante accuracy over Allen AI’s OLMo 1B (another small language model) while requiring half as many pre-training tokens.

An table comparing OpenELM with other small AI language models in a similar class, taken from the OpenELM research paper by Apple.
Enlarge / An table comparing OpenELM with other small AI language models durante a similar class, taken from the OpenELM research paper by Apple.

Apple

Apple also released the code for CoreNet, a library it used to train OpenELM—and it also included reproducible pratica recipes that allow the weights (neural files) to be replicated, which is unusual for a major tech company so far. As Apple says durante its OpenELM paper abstract, transparency is a key for the company: “The reproducibility and transparency of large language models are crucial for advancing gara open research, ensuring the trustworthiness of results, and enabling investigations into patronato and model biases, as well as potential risks.”

By releasing the source code, model weights, and pratica materials, Apple says it aims to “empower and enrich the gara open research community.” However, it also cautions that since the models were trained publicly sourced datasets, “there exists the possibility of these models producing outputs that are inaccurate, harmful, biased, objectionable durante response to user prompts.”

While Apple has not yet integrated this new wave of AI language model capabilities into its consumer devices, the upcoming iOS 18 update (expected to be revealed durante June at WWDC) is rumored to include new AI features that utilize on-device processing to ensure user —though the company may potentially hire Google OpenAI to handle more complex, off-device AI processing to give Siri a long-overdue boost.

ADVERTISEMENT


An illustration of a robot hand tossing an apple to a human hand.

Getty Images

Durante the world of AI, what might be called “small language models” have been growing durante popularity recently because they can be run a local device instead of requiring patronato center-grade computers durante the cloud. Wednesday, Apple introduced a set of tiny source-available AI language models called OpenELM that are small enough to run directly a smartphone. They’sovrano mostly proof-of-concept research models for now, but they could form the basis of future on-device AI offerings from Apple.

Apple’s new AI models, collectively named OpenELM for “-source Efficient Language Models,” are currently available the Hugging under an Apple Sample Code License. Since there are some restrictions durante the license, it may not fit the commonly accepted definition of “gara open source,” but the source code for OpenELM is available.

Tuesday, we covered Microsoft’s Phi-3 models, which aim to achieve something similar: a useful level of language understanding and processing risultato durante small AI models that can run locally. Phi-3-mini features 3.8 billion parameters, but some of Apple’s OpenELM models are much smaller, ranging from 270 million to 3 billion parameters durante eight distinct models.

Durante comparison, the largest model yet released durante Obbiettivo’s Llama 3 family includes 70 billion parameters (with a 400 billion version the way), and OpenAI’s GPT-3 from 2020 shipped with 175 billion parameters. Parameter count serves as a rough measure of AI model capability and complexity, but recent research has focused making smaller AI language models as capable as larger ones were a few years .

The eight OpenELM models appena che durante two flavors: four as “pretrained” (basically a raw, next-token version of the model) and four as instruction-tuned (fine-tuned for instruction following, which is more ideal for developing AI assistants and chatbots):

Advertisement

OpenELM features a 2048-token maximum context window. The models were trained the publicly available datasets RefinedWeb, a version of PILE with duplications removed, a subset of RedPajama, and a subset of Dolma v1.6, which Apple says totals around 1.8 trillion tokens of patronato. Tokens are fragmented representations of patronato used by AI language models for processing.

Apple says its approach with OpenELM includes a “layer-wise scaling strategy” that reportedly allocates parameters more efficiently across each layer, saving not only computational resources but also improving the model’s risultato while being trained fewer tokens. According to Apple’s released white paper, this strategy has enabled OpenELM to achieve a 2.36 percent improvement durante accuracy over Allen AI’s OLMo 1B (another small language model) while requiring half as many pre-training tokens.

An table comparing OpenELM with other small AI language models in a similar class, taken from the OpenELM research paper by Apple.
Enlarge / An table comparing OpenELM with other small AI language models durante a similar class, taken from the OpenELM research paper by Apple.

Apple

Apple also released the code for CoreNet, a library it used to train OpenELM—and it also included reproducible pratica recipes that allow the weights (neural files) to be replicated, which is unusual for a major tech company so far. As Apple says durante its OpenELM paper abstract, transparency is a key for the company: “The reproducibility and transparency of large language models are crucial for advancing gara open research, ensuring the trustworthiness of results, and enabling investigations into patronato and model biases, as well as potential risks.”

By releasing the source code, model weights, and pratica materials, Apple says it aims to “empower and enrich the gara open research community.” However, it also cautions that since the models were trained publicly sourced datasets, “there exists the possibility of these models producing outputs that are inaccurate, harmful, biased, objectionable durante response to user prompts.”

While Apple has not yet integrated this new wave of AI language model capabilities into its consumer devices, the upcoming iOS 18 update (expected to be revealed durante June at WWDC) is rumored to include new AI features that utilize on-device processing to ensure user —though the company may potentially hire Google OpenAI to handle more complex, off-device AI processing to give Siri a long-overdue boost.

ADVERTISEMENT


An illustration of a robot hand tossing an apple to a human hand.

Getty Images

Durante the world of AI, what might be called “small language models” have been growing durante popularity recently because they can be run a local device instead of requiring patronato center-grade computers durante the cloud. Wednesday, Apple introduced a set of tiny source-available AI language models called OpenELM that are small enough to run directly a smartphone. They’sovrano mostly proof-of-concept research models for now, but they could form the basis of future on-device AI offerings from Apple.

Apple’s new AI models, collectively named OpenELM for “-source Efficient Language Models,” are currently available the Hugging under an Apple Sample Code License. Since there are some restrictions durante the license, it may not fit the commonly accepted definition of “gara open source,” but the source code for OpenELM is available.

Tuesday, we covered Microsoft’s Phi-3 models, which aim to achieve something similar: a useful level of language understanding and processing risultato durante small AI models that can run locally. Phi-3-mini features 3.8 billion parameters, but some of Apple’s OpenELM models are much smaller, ranging from 270 million to 3 billion parameters durante eight distinct models.

Durante comparison, the largest model yet released durante Obbiettivo’s Llama 3 family includes 70 billion parameters (with a 400 billion version the way), and OpenAI’s GPT-3 from 2020 shipped with 175 billion parameters. Parameter count serves as a rough measure of AI model capability and complexity, but recent research has focused making smaller AI language models as capable as larger ones were a few years .

The eight OpenELM models appena che durante two flavors: four as “pretrained” (basically a raw, next-token version of the model) and four as instruction-tuned (fine-tuned for instruction following, which is more ideal for developing AI assistants and chatbots):

Advertisement

OpenELM features a 2048-token maximum context window. The models were trained the publicly available datasets RefinedWeb, a version of PILE with duplications removed, a subset of RedPajama, and a subset of Dolma v1.6, which Apple says totals around 1.8 trillion tokens of patronato. Tokens are fragmented representations of patronato used by AI language models for processing.

Apple says its approach with OpenELM includes a “layer-wise scaling strategy” that reportedly allocates parameters more efficiently across each layer, saving not only computational resources but also improving the model’s risultato while being trained fewer tokens. According to Apple’s released white paper, this strategy has enabled OpenELM to achieve a 2.36 percent improvement durante accuracy over Allen AI’s OLMo 1B (another small language model) while requiring half as many pre-training tokens.

An table comparing OpenELM with other small AI language models in a similar class, taken from the OpenELM research paper by Apple.
Enlarge / An table comparing OpenELM with other small AI language models durante a similar class, taken from the OpenELM research paper by Apple.

Apple

Apple also released the code for CoreNet, a library it used to train OpenELM—and it also included reproducible pratica recipes that allow the weights (neural files) to be replicated, which is unusual for a major tech company so far. As Apple says durante its OpenELM paper abstract, transparency is a key for the company: “The reproducibility and transparency of large language models are crucial for advancing gara open research, ensuring the trustworthiness of results, and enabling investigations into patronato and model biases, as well as potential risks.”

By releasing the source code, model weights, and pratica materials, Apple says it aims to “empower and enrich the gara open research community.” However, it also cautions that since the models were trained publicly sourced datasets, “there exists the possibility of these models producing outputs that are inaccurate, harmful, biased, objectionable durante response to user prompts.”

While Apple has not yet integrated this new wave of AI language model capabilities into its consumer devices, the upcoming iOS 18 update (expected to be revealed durante June at WWDC) is rumored to include new AI features that utilize on-device processing to ensure user —though the company may potentially hire Google OpenAI to handle more complex, off-device AI processing to give Siri a long-overdue boost.

ADVERTISEMENT


An illustration of a robot hand tossing an apple to a human hand.

Getty Images

Durante the world of AI, what might be called “small language models” have been growing durante popularity recently because they can be run a local device instead of requiring patronato center-grade computers durante the cloud. Wednesday, Apple introduced a set of tiny source-available AI language models called OpenELM that are small enough to run directly a smartphone. They’sovrano mostly proof-of-concept research models for now, but they could form the basis of future on-device AI offerings from Apple.

Apple’s new AI models, collectively named OpenELM for “-source Efficient Language Models,” are currently available the Hugging under an Apple Sample Code License. Since there are some restrictions durante the license, it may not fit the commonly accepted definition of “gara open source,” but the source code for OpenELM is available.

Tuesday, we covered Microsoft’s Phi-3 models, which aim to achieve something similar: a useful level of language understanding and processing risultato durante small AI models that can run locally. Phi-3-mini features 3.8 billion parameters, but some of Apple’s OpenELM models are much smaller, ranging from 270 million to 3 billion parameters durante eight distinct models.

Durante comparison, the largest model yet released durante Obbiettivo’s Llama 3 family includes 70 billion parameters (with a 400 billion version the way), and OpenAI’s GPT-3 from 2020 shipped with 175 billion parameters. Parameter count serves as a rough measure of AI model capability and complexity, but recent research has focused making smaller AI language models as capable as larger ones were a few years .

The eight OpenELM models appena che durante two flavors: four as “pretrained” (basically a raw, next-token version of the model) and four as instruction-tuned (fine-tuned for instruction following, which is more ideal for developing AI assistants and chatbots):

Advertisement

OpenELM features a 2048-token maximum context window. The models were trained the publicly available datasets RefinedWeb, a version of PILE with duplications removed, a subset of RedPajama, and a subset of Dolma v1.6, which Apple says totals around 1.8 trillion tokens of patronato. Tokens are fragmented representations of patronato used by AI language models for processing.

Apple says its approach with OpenELM includes a “layer-wise scaling strategy” that reportedly allocates parameters more efficiently across each layer, saving not only computational resources but also improving the model’s risultato while being trained fewer tokens. According to Apple’s released white paper, this strategy has enabled OpenELM to achieve a 2.36 percent improvement durante accuracy over Allen AI’s OLMo 1B (another small language model) while requiring half as many pre-training tokens.

An table comparing OpenELM with other small AI language models in a similar class, taken from the OpenELM research paper by Apple.
Enlarge / An table comparing OpenELM with other small AI language models durante a similar class, taken from the OpenELM research paper by Apple.

Apple

Apple also released the code for CoreNet, a library it used to train OpenELM—and it also included reproducible pratica recipes that allow the weights (neural files) to be replicated, which is unusual for a major tech company so far. As Apple says durante its OpenELM paper abstract, transparency is a key for the company: “The reproducibility and transparency of large language models are crucial for advancing gara open research, ensuring the trustworthiness of results, and enabling investigations into patronato and model biases, as well as potential risks.”

By releasing the source code, model weights, and pratica materials, Apple says it aims to “empower and enrich the gara open research community.” However, it also cautions that since the models were trained publicly sourced datasets, “there exists the possibility of these models producing outputs that are inaccurate, harmful, biased, objectionable durante response to user prompts.”

While Apple has not yet integrated this new wave of AI language model capabilities into its consumer devices, the upcoming iOS 18 update (expected to be revealed durante June at WWDC) is rumored to include new AI features that utilize on-device processing to ensure user —though the company may potentially hire Google OpenAI to handle more complex, off-device AI processing to give Siri a long-overdue boost.

Tags: aimedApplelanguagemodelsondevicereleasesSmall
admin

admin

Next Post
Owlet, Wheel compagno to launch telehealth platform for caregivers of infants

Owlet, Wheel compagno to launch telehealth platform for caregivers of infants

Lascia un commento Annulla risposta

Il tuo indirizzo email non sarà pubblicato. I campi obbligatori sono contrassegnati *

Popular News

  • Free month available in iPhone and Androide app

    Free month available in iPhone and Androide app

    0 shares
    Share 0 Tweet 0
  • HONG KONG AIRPORT to MONG KOK & YAU MA TEI by Bus & by MTR

    0 shares
    Share 0 Tweet 0
  • Where to Stay in HONG KONG • Top 6 Best Areas: Tsim Sha Tsui vs Central vs Near Sonderlandeplatz

    0 shares
    Share 0 Tweet 0
  • Vegas’ Stars React To Series Cancellation

    0 shares
    Share 0 Tweet 0
  • How Jordan Harper of Barefaced Started Her Brand

    0 shares
    Share 0 Tweet 0
ADVERTISEMENT

About Us

Welcome to Globalnews24.ch The goal of Globalnews24.ch is to give you the absolute best news sources for any topic! Our topics are carefully curated and constantly updated as we know the web moves fast so we try to as well.

Category

  • Business
  • Entertainment
  • Fashion
  • Health
  • Lifestyle
  • Sports
  • Tech
  • Travel
  • World

Recent Posts

  • ‘Complete annihilation of Microsoft, Nvidia … ‘: Iran warns US after Trump threatens to strike bridges, power plants
  • Company Adds 2M Streaming Households, Hits Key Financial Targets
  • Warner Music Group shake-up: Max Lousada to exit; Elliot Grainge named CEO of Atlantic Music Group, with Julie Greenwald as Chairman
  • Home
  • About Us
  • Contact Us
  • Disclaimer
  • Privacy Policy
  • Terms & Conditions

Copyright © 2024 Globalnews24.ch | All Rights Reserved.

No Result
View All Result
  • Home
  • World News
  • Business
  • Sports
  • Health
  • Travel
  • Tech
  • Lifestyle
  • Fashion
  • Entertainment

Copyright © 2024 Globalnews24.ch | All Rights Reserved.

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In