The tech industry is finally listening to calls for more inclusive AI—but at what cost? As Howard University and Google launch Project Elevate Black Voices, the spotlight is now on collecting African American English (AAE) data to improve AI systems. While this is a much-needed move to correct decades of bias in tech tools, it also raises a crucial question: Who’s protecting Black voices in AI as they become a powerful asset for Big Tech?
The partnership between Howard University and Google is collecting over 600 hours of speech data across 32 states to represent the wide spectrum of African American English. This data can improve tools like Zoom captions, AI notetakers, and transcription software that often fail to understand AAE. From healthcare to legal sectors, these upgrades could make workplace tech more accurate and inclusive—especially for Black professionals who’ve had to “code-switch” just to be understood.
Despite its promise, many remain skeptical—and for good reason. The tech industry has a long record of extracting from Black communities without credit or compensation. Black creators have sparked viral trends only to be erased, while algorithms disproportionately flag their content or misidentify them in surveillance systems. Critics fear this dataset could be misused for digital Blackface, commercial AI avatars, or surveillance tools that further harm the very people it's meant to help.
Howard has committed to licensing the dataset exclusively to HBCUs—for now. But questions remain: Who will access this data next? Will the community benefit financially or culturally? Even within Black-led spaces, harm can still happen. As Audre Lorde warned, oppression can be replicated from within. To truly protect Black voices in AI, we must build systems rooted in community power—not just corporate interest—and ensure that AI tools are built with Black communities, not just about them.
There’s no doubt that the Howard African American English Dataset can help make AI better, fairer, and more inclusive. But we must remain cautious. Protecting Black voices in AI means setting boundaries, demanding transparency, and ensuring that Black researchers, workers, and communities are not only represented—but respected, compensated, and in control. Tech may be evolving, but without accountability, history will repeat itself—this time coded into the machines we use every day.
Semasocial is where real people connect, grow, and belong.
We’re more than just a social platform — we’re a space for meaningful conversations, finding jobs, sharing ideas, and building supportive communities. Whether you're looking to join groups that match your interests, discover new opportunities, post your thoughts, or learn from others — Semasocial brings it all together in one simple experience.
From blogs and jobs to events and daily chats, Semasocial helps you stay connected to what truly matters.