amberSearch

Our Open Source Contributions

Over the years, we have published parts of our code and our AI models. We have summarised these contributions on this page.

Multilingual Passage Reranking Model

High quality, German training data set

  • High-quality data sets are required to train AI models and avoid certain biases.
  • Together with Philipp May, T-Systems, we have defined this training dataset to train our LLMs.
  • When we train AI models, we do not use customer data, but general, public data – such as from this training dataset.

German Electra Model

  • Together with Philip May, T-Systems, we have published an improved Electra model, which was trained with 1.5 million steps.
  • At the time of publication in August 2020, it was briefly the most efficient German-language LLM.
  • Of course, it is now far from state of the art. But it shows how early we started working with large language models and that we have own expertise in this technology.

You are currently viewing a placeholder content from YouTube. To access the actual content, click the button below. Please note that doing so will share data with third-party providers.

More Information