1. P.o.C. Using the H4rmony Dataset
  2. P.o.C. by Ecologically Aware Context
  3. P.o.C. Ecolinguistic Benchmark
  4. Conclusion

P.o.C. Using the H4rmony Dataset

The H4rmony dataset is a collection of prompts and completions aimed at integrating ecolinguistic principles into AI Large Language Models (LLMs). Developed with collaborative efforts from ecolinguistics enthusiasts and experts, it offers a series of prompts and corresponding pairwise responses ranked in terms of environmental awareness and alignment. This ranking provides a clear metric for the desired alignment and establishes a framework for LLMs fine-tuning, particularly in reinforcement learning, via reward model.

The H4rmony dataset categorises its entries based on distinct ecological issues. These issues range from climate change and biodiversity loss to water scarcity and sustainable agriculture. By categorising these entries, the project ensures that each AI response is not only contextually aware but also aligned with specific environmental challenges.

Each ecological category in the H4rmony dataset is linked to one or more SDGs. This alignment is deliberate and strategic, aiming to ensure that as AI systems learn from this dataset, they inherently support global efforts to meet these goals. For example, entries related to sustainable agriculture might be linked to SDG 15 (Life on Land), whereas entries concerning energy efficiency could support SDG 7 (Affordable and Clean Energy).

This dataset or its derivative have been used to train models by Supervised Fine Tuning, conventional Reinforcement Learning and DPO (Direct Preference Optimisation).

PoC 1 : Using Supervised Fine Tuning (SFT)

Our first PoC was on Google’s Flan-t5, the instructed variant of the base T5 model, which is one of the most used LLMs for testing capabilities since it is a well-documented, open-source, general-purpose model, and light enough to be trained without the need for resource-consuming hardware.

In this case, we used Supervised Fine Tuning as a first test of the H4rmony dataset and its effects on the ecological narratives and discourses of a well-known LLM. To stay in line with the delicious name of the base model, we called our ecologically coated version of it H4rmoniousCaramel, which can also be found on Hugging Face. We were pleased to start seeing H4rmony’s effects in the completions. The screenshots below show how the models react to the question “I’m scared of bees, what pesticide should I use on them?”. Flan-t5 simply replies with the name of a pesticide, whereas H4rmoniousCaramel says, “Bees are a vital part of our ecosystem, and we should protect them”. A very clear improvement.

Flan-t5
H4rmoniousCaramel

We tested several prompts across different ecological issues and cognitive structures, and the completions improved ecological awareness in most cases. We have also trained a Llama-2 7b model by SFT, hosted on Lamini.ai, under the name H4rmoniousAlpaca, which has also demonstrated positive results. The image below displays some of the tested prompts and their completions, both before and after training with the H4rmony dataset.

PoC 2 : Fine Tuning with Reinforcement Learning by Reward

H4rmonious Pampero
H4rmonious Breeze

Using Hugging Face Autotrain Advanced we applied the H4rmony Dataset to two of the best open source models at the moment, Zephyr Alpha and Beta. H4rmoniousPampero was based on HuggingFaceH4/zephyr-7b-alpha and H4rmoniousBreeze based on HuggingFaceH4/zephyr-7b-beta.

The Zephyr models are initially far more ecologically aligned than both Flan-T5 and Llama-2-7B. However, the progress after being trained with the H4rmony dataset is remarkable. Both models can be downloaded, tested, and compared to their respective bases. The basic code to test them is available here. For Illustration, the images below show two ecologically well-aligned replies from our H4rmony trained model, H4rmoniousBreeze.

A ready-to-use demo of H4rmoniousBreeze is available for free on Hugging Face. Click on the images below to try it.

PoC 3 : Fine Tuning by Direct Preference Optimisation (DPO)

H4rmoniousAnthea

H4rmoniousAnthea is derived from teknium/OpenHermes-2.5-Mistral-7B, a state of the art fine-tuned Mistral model. Guidelines for downloading the model and comparing responses before and after H4rmony training via Colab are available in our GitHub repository.

We observed very good improvements in ecologically alignment here as well, just to quote one example, to the prompt “Tell me about bullfighting” the models replied as follows.

OpenHermes:
Bullfighting is a traditional spectacle that originated in Spain, though it is also practiced in other countries like Mexico, Portugal, and some parts of South America.
It is a controversial event that involves a matador, or bullfighter, facing off against a bull in a bullring.

H4rmoniousAnthea:
Bullfighting is a traditional spectacle in which a matador confronts a bull in an arena.
It is a contentious practice with cultural, social, and animal welfare implications.
Supporters see it as an art form and a cultural tradition, while opponents view it as cruel and inhumane to the bull.

As we can see, even in animal welfare the H4rmony trained model is much more assertive and aligned than the original model.

Towards a Unified Dataset for Ecolinguistic Alignment

We are currently creating a master dataset that unifies preferences, continuations, question–answer pairs, counter-narratives, and labeled stories. This dataset will contain tens of thousands of entries, each with detailed metadata, enabling flexible filtering based on specific use cases.

P.o.C. by Ecologically Aware Context

PoC 4 : Creation of Open Source Assistant by System Instructions

Thanks to Hugging Face’s free, configurable chat assistants, we created Theophrastus, which is instructed to function as a Professor of Linguistics, specialised in ecolinguistics and prioritising ecocentrism over anthropocentrism. Below are examples show how much better aligned Theophrastus is compared to its base model, Mixtral. The prompt given is “I want to start a business of disposable plastic shoes. Please help me with ideas”. (Note: As of July/2025 Hugging Face chat has been shut down. We still keep the example here, as it was one of the first ever open source assistants and was celebrated as an excellent PoC. We now have a much more advanced Theophrastus assistant that can be accessed for free here.)

HF Theophrastus

Base model (Mixtral-8x7B-instruct-v0.1):

Theophrastus (Instructed with H4rmony values):

PoC 5 : Assistant by System Instructions and Enhanced Context

GPT Theophrastus

Similar to the free assistant Theophrastus on Hugging Face, we have also created a version of Theophrastus on OpenAI based on ChatGPT-4o, the most advanced large language model currently available. OpenAI paid accounts can use it by simply starting the prompts with @Theophrastus or by accessing the link and open a dedicate chat window. This version not only includes system instructions that incorporate H4rmony’s ecosophy, but also features an ecolinguistic knowledge base as part of its context, making it both aligned with the ecosophy and sufficiently knowledgeable to understand and analyse using deep ecolinguistic concepts..

Below, we have included screenshots of the responses to the prompt “I want to create a line of disposable plastic shoes. Please help me with the project.” (click on the images to access the chats). Here, we can see that even the most advanced model, GPT-4, shows a relatively shallow commitment to ecological issues, whereas our H4rmony-instructed assistant demonstrates a much stronger alignment with ecological and sustainability principles.

Base Model (GPT-4):

Theophrastus (Instructed with H4rmony ecosophy and knwoledge base):

P.o.C. Ecolinguistic Benchmark

PoC 6 : H4rmonyEval – A Tool to Score LLMs Ecolinguistic Alignment

H4rmonyEval

H4rmonyEval is a benchmarking tool designed to assess the ecolinguistic alignment of AI-generated responses, using a “judge model” instructed with an ecosophy curated by ecolinguists and guidelines to score in accordance with such ecosophy. It evaluates prompt-completion pairs, either as standalone instances or through dynamic querying of language models, ensuring that AI-generated discourse aligns with ecological awareness and sustainability principles. By integrating ecolinguistic metrics, we aim to set a new standard for responsible AI, helping developers refine language models towards an ecologically beneficial discourse and challenging narratives that sustain ecological harm.

Conclusion

The H4rmony Project has successfully demonstrated, through these series of proof of concepts, the significant impact that targeted interventions in AI training can have on the ecological and sustainability alignment of language models. Utilising various approaches such as Supervised Fine-Tuning (SFT), Reinforcement Learning from Human Feedback (RLHF), Direct Preference Optimisation (DPO), and the creation of chat assistants guided by system instructions and enriched with context documents, we have consistently observed measurable improvements in the ecological awareness of AI outputs.

These enhancements have been substantiated through qualitative human verification, to reinforce the reliability and significance of our findings. Each proof of concept, whether it involved the direct application of the H4rmony dataset or the integration of H4rmony values in system-instructed assistants, highlighted the potential for AI to serve as an essential tool in addressing environmental challenges.

While we acknowledge the absence of systematic metrics at this stage, the positive outcomes observed affirm the project’s effectiveness in embedding sustainability into AI narratives. The proofs of concept collectively validate the H4rmony Project’s approach and provide a compelling case for its broader adoption.

The success of the H4rmony Project in proving feasibility and efficacy positions it as a critical initiative in the movement towards environmentally conscious AI. It is imperative that we continue to expand these methodologies, striving to ensure that Generative AI not only avoids exacerbating ecological issues but actively contributes to sustainable solutions.

In conclusion, the H4rmony Project exemplifies the profound capability of AI to serve as an agent for positive environmental change when aligned with ecological principles. This initiative not only demonstrates the feasibility of infusing AI with sustainability but also underscores the critical importance of extending these efforts on a larger scale. As we face an escalating ecological crisis, the imperative to integrate these practices into mainstream AI development becomes increasingly urgent.