• Skip to main content
  • Skip to footer
  • Home

The May 13 Group

the next day for evaluation

  • Get Involved
  • Our Work
  • About Us
You are here: Home / Archives for allblogs

allblogs

May 28 2023

AEA Conference 2022

Preconference Workshop: Transformative Mixed Methods: Supporting Equity & Justice by Donna Mertens

A key reason that I decided to take this pre-conference workshop was because I wanted to learn from Donna Mertens. I really like her writing and wanted to have a chance to learn from her in person. She did not disappoint! While I didn’t find there was much about mixed methods, per se, there was a lot about transformation, equity, and justice. Here are some things I learned/re-learned:

  • In France there is a law against the government collecting data on race. It comes from WWII when government data on Jewish people facilitated the ability to send Jewish people to concentration camps.
  • Ethics is the start of every decision we make in evaluation.
  • If you are not challenging oppressive structures, you are complicit in the status quo.
  • You can challenge your client/commissioner – e.g., if they ask for a survey for a summative evaluation, you can ask them if that’s really going to be transformative.
  • mixed methods – what is the synergy between the quant and the qual? what do you gain by bringing quant and qual in dialogue with each other?
  • transformative paradigm
    • axiology (i.e., nature of ethics & values) – culturally responsive, promotes social/environmental/economic justice and human rights, address inequities, reciprocity (what do you leave the community so they can sustain the change when the evaluator leaves?), resilience, interconnectedness (living & non-living), relationships
    • ontology (i.e., nature of reality) – reality is multi-faceted, historically situated, consequences of privilege
    • epistemology (i.e., the nature of knowledge & relationship between knower and that which would be known) – interactive, trust, coalition building
    • methodology (i.e., nature of systematic inquiry) – transformative, dialogic, culturally responsive, mixed methods, policy change as part of methodology
  • transformative mixed methods design:
    • Build relationships
      • often historical experiences of research/evaluation that are extractive and oppressive; researchers need to earn trust
      • identifying existing community actions groups and understand the history of their efforts; identify formal & informal leaders; identifying community needs/gaps/strengths/assets
    • Contextual analysis
      • cultural, historical, political, environmental, legislative, power mapping
      • policy analysis (what’s written and unwritten; what’s written by not enacted)
    • Pilot interventions
      • collect data, make mid-course corrections
    • Implement intevetnion
      • collect data for proces evaluation
      • collect data on unintended/unanticipated outcomes
    • Determine effectiveness
      • outcome evalaution
    • Use findings for transformative purposes
      • include in contract the importance of working with the community – from relationship building at the start all the way through to sharing the findings at the end
      • if the community is involved throughout the evaluation, they will already know the findings and will not need to wait for the final report to find out (also, reporting findings along the way will make sure you are reporting data back to community in a reasonable time)
  • You can say that you’ll work with an expected goal of reducing inequities and increased justice and that you’ll work in respective ways; you can’t guarantee that you’ll make things better and can’t guarantee you won’t cause harm, because we don’t know what will happen
  • http://transformativeresearchandevaluation.com/

Opening Plenary: Re(Shaping) Evaluation: Decolonization, New Actors, & Digital Data. Edgar Villanueva interviewed by Nicky Bowman.

Villaneuva wrote the book Decolonizing Wealth. I will admit that I have this book in my big pile of books to read, but hadn’t got around to reading it! After hearing this keynote, I’m even more excited to read it. Here are some things he said during the keynote that resonated with me:

  • we learned the names of the colonizers’ ships (the Nina, the Pinta, the Santa Maria, the Mayflower), but not the names of the Indigenous lands and people
  • colonization is like a virus that wipes out anything that is not like the dominant culture
  • the US is working on Truth & Reconciliation legislation re: Indian Boarding Schools
  • none of us has ever lived in a world that wasn’t actively being colonized. It can be violent and it can be subtle.
  • we can’t collectively heal without acknowledging how we got here
  • we need to change 4 things:
    1. people: more diversity of perspectives in leadership
    2. resources: who has them and who makes decisions about them. who has the microphone.
    3. stories: need to shift away from the deficit mindset, see the strengths
    4. rules: spoken & unspoken policies, need more equitable policies, but also need to become aware of and change the unspoken rules that limit our work

Concurrent Session: Walking the talk: Bringing Ontological Thinking into Evaluation Practice by Jennifer Billman and Eric Einspruch

Journal article: Framing Evaluation in Reality: An Introduction to Ontologically Integrative Evaluation

Thursday Plenary: Co-creation of Strategic Evaluations to Shift Power Moderator: Ayesha Boyce Speakers: Elizabeth Taylor-Schiro/Biidabinikwe, Gabriela Garcia, Melanie Kawano-Chiu, and Subarna Mathes 

Here are some things that the panelists said that resonated with me:

  • Ayesha Boyce:
    • equity is context-specific
  • Gabriela Garcia:
    • equity is not enough. The next step is collective liberation
    • At Beyond, they use a culturally-responsive evaluation framework, start all evaluations in a visioning session, ensure the evaluation is grounded in community values
  • Elizabeth Taylor-Schiro/Biidabinikwe:
    • communities striving for collective liberation don’t have power and that’s the problem. Power is needed to draw on their strengths, move toward sustainability and self-determination
    • it should be the community leading the evaluation, supported by evaluators, rather than ‘co-creating’ the evaluation
  • Melanie Kawano-Chiu:
    • whoever funds the evaluation gets to make the most decisions – that’s a bias we hold
    • ableism = there is a “norm” and if you fall outside of it, you aren’t good enough
    • “Nothing about us without us” comes from the South African disability community
    • Disability Rights Advocacy Fund formed after the UN Convention on Rights of People with Disabilities
  • Subarna Mathes:
    • if rigour = degree of confidence that the program has led to an outcome [different than rigour in the post-positivist sense]
    • we need to push against the view of rigour that is narrowly defined, that prioritizes a worldview of “one reality” or “objectivity”

Concurrent session: Interactive tool to promote responsible use and understanding of culturally responsive and equity-focused evaluation by Blanca Guillen-Woods, Felisa Gonzales, Katrina Bledsoe, Kantahyanee Murray

  • https://slp4i.com/the-eval-matrix/ is an online tool that helps you to choose from various different equity-focused/culturally-responsive evaluation approaches
  • 7 key principles, 3 focus areas (individual, interpersonal, structural levels)
  • This tool is really cool and I’m definitely going to share it with my students, as they often ask how to choose an approach (or approaches) when designing an evaluation

Concurrent Session: Design Sprint: How Researchers Can Share Power with Communities Involved in Evaluations by Gloriela Iguina-Colón and Brit Henderson

These presenters took us through a workshop on power sharing. Here are some things that they talked about that resonated with me:

  • power is often thought of in the sense of authority, control – power over other people or things
  • MLK descrbied power as “Power properly understood is nothing but the ability to achieve purpose. It is the strength required to bring about social, political, and economic change.”
  • can have power with (collaborate with others to find common ground), power to (believe in people’s ability to shape their own lives), and power within [which I didn’t catch the meaning of in my notes so I just Googled it and found this: “the sense of confidence, dignity and self-esteem that comes from gaining awareness of one’s situation and realizing the possibility of doing something about it.”]
  • power levers:
    • resources
    • access
    • opportunities
  • power sharing = recognizing the power levers that you have and actively choosing to leverage these to build collective strength
  • positionality – “how our social identities and experiences influence the choices we make in the research process and how those factors shape the way others see us and give us power and/or insight in a specific research context.”
    • consider experiences (interactions with the topic; lived experience of the topic), social identities (it’s context-dependent which are valued or not valued), perspectives (about the topic; understanding systems of oppression); identifying these can provide helpful insights (e.g., when you share an identity with participants) and biases (e.g., when you don’t share an identity (or an intersection of identities) and have assumptions/biases)
      • in addition to individual positionality, think about team positionality
  • reflectivity – “an attitude of attending systematically to the context of knowledge construction, especially to the effect of the researcher, at every step of the research practice.
    • examination, attitude, process related to the topic; not just about identifying these things, but also what insights this will give me and where I might have knowledge gaps
  • opportunity spaces: “points in the [evaluation] process during which you can apply power levers to facilitate meaningful participation among and share decision-making power with
    • each of the steps of the evaluation process is an opportunity for meaningful participation: evaluation design, data collection, data analysis, interpretation of results, and dissemination
  • facilitation – provide enough structure so everyone can be heard; be mindful of different views of evaluation
  • when considering the key people/groups in an evaluation, ask:
    • who has the most power/privilege in this context?
    • who will be most impacted by the evaluation?

Concurrent Session: Ethics for Evaluation: Can We Go Beyond Doing No Harm to Tackle Bad and Do Good? by Penny Hawkins, Donna Mertens, and Tessie Catsambas

  • Ethics for Evaluation: Beyond “doing no harm” to “tackling bad” and “doing good”. Edited By Rob D. van den Berg, Penny Hawkins, Nicoletta Stame

Concurrent Session: Equitable Evaluation Discussion Guide by Maggie Jones, Natasha Arora, and Elena Kuo

  • Centre for Community Health & Evaluation at Kaiser Permenated (Seattle) [Seemed quite similar to CHEOS)
  • equity-focused conversation about the evaluation design with someone from their organization who is not part of the project to get a different perspective
  • they created a guide that includes pre-work before the meeting, them a meeting where you do a consultation with reviewers
  • helped them to think from multiple angles (not just “what’s in the RFP?”)
  • helped them to discuss assumptions and implications
  • articulate what they can and cannot do to address equity
    • might not be able to do something in the current evaluation, but if you don’t identify ideas, won’t ever do them – so may be something to put in the next proposal if it’s too late to do it in this proposal
  • they tell funders that the EDI reviews is part of their process (i.e., we will develop the plan, put it through the EDI review and may come back to the funder with new ideas)
  • ultimately would like to have a systematic follow up process where people will document what they do (trying to document changes that happened due to the EDI review process) to build evidence if this process makes a difference

Concurrent Session: Identifying Gaps in the Research on Professionalizing Evaluation: What Do We Need? by Amanda Sutter, Esther Nolton, Rebecca Teasdale, Rachael Kenney, Dana Wanzer

Concurrent Session: Creative Practices for Evaluators by Chantal Hoff & Susan Putnins

  • reminded me a lot of Jennica & May from ANDImplementation

Concurrent Session: Who Are We? Studies on Evaluators’ Values, Ethics and Ontologies by John M. LaVelle, Michael Morris, Clayton Stephenson, Scott I Donaldson, Justin Hacket, Paidamoyo Chikate, Jennifer Billman

  • VOPEs have ethics, standards, and competencies, but we as evaluator interpret them through our own lenses
  • values = a set of goals and motivations that serve as a guiding group of principles, affect decisions/attitudes/behaviours, come from many sources, influence our practice

Concurrent Session: Mapping Distinctions in the Implementation of Learning Health System (LHS) by Anna Perry & Dough Easterling

  • from National Academy of Medicine, but concept is too high level and ambiguous to guide the actual work of becoming a LHS
  • in the US, electronic health records adopted in early 2000s, Afforable Care Act required the use of data to inform the health system
  • Academci Helath Centres not early adopters of LHS because they were focused on research to build knowledge vs. continuous improvement type stuff
  • hypothesis is that LHS is supposed to improve patient care, patient outcomes, and staff satisfaction (since they are more engaged)

Concurrent Session: Who are We? Studies of Evaluator Beliefs, Identify, and Ethics by Rachel Kenney, Bianca Montrosse-Moorhead, Amanda Sutter, Christina Peterson, Rachel Ladd, Betty Onyura, Abigail Fisher, Qian Wu, Shrutikaa Rajkumar, Sarick Chapagain, Judith Nassuna, Latika Nirula

  • Ladd & Peterson discussed consensual qualitative analysis
  • Tin Vo presented on behalf of Betty Onyura, who was not able to attend. Talked about how the commodification of evaluation work is in tension with trying to support equity and social justice
Untitled
  • an audience member suggested the word “constituent” instead of “stakeholder” [as a lot of us are trying to find a word to replace “stakeholder”]

Concurrent Session: Playing with Uncertainty: Leaning into Improv for Effective Evaluation by Daniel Tsin, Libby Smith, and Tiffany Tovey

  • improv as reflection-in-action
  • improv as a mindset – every idea matters
  • thinking on your feet, using a different part of your brain, building on ideas, chance to be brave – can all be useful in evaluation
  • activity: Zip Zap Zop – toss a ball and say “zip,” “zap”, “zop” in that order and when someone drops the ball, we all cheer “woop!”
    • a chance to experience failure and turn it into a celebration
    • shared experience of a group
    • a plan for when we messed up
    • have to pay attention the whole time – not planning what to do, but being present, acknowledging what is being said/done
    • facilitator is not in control
  • activity: Yes, and…
    • “and” is generative, while “but” feels more like you are shutting someone down
    • you will notice “but” in every day life when you could have used an “and”
    • sometimes you want to be generative and sometimes you want to prioritize (e.g., don’t want to keep “and”ing when building a program ToC and end up with trying to do everything).
  • Adrienne Marie Brown’s Emergent Strategy

Concurrent Session: What Should I Do? Examining Uncertainty, Decisions Points, and Pushback in Evaluation Practice by Rebecca Tesadale, Tiffany Tovey, Grettel Arias Orozco, Julianne Zemaitis, Onyinyechukwu Onwuka, Cherie Avent, Christina Peterson, Allison Ricket, Mandy White, Kelli Schoen, Daniel Kloepfer, Natalie Wilson

  • evaluations require interpersonal skills, but it’s not taught in evaluation courses or in evaluation texts
  • it’s a human tendency to be defensive and as a conversation proceeds, defensiveness will increase as what a person hears can become a distortion of what the message was
  • Kahlke, 2014 – Generic Qualitative Approaches: Pitfalls and Benefits of Methodological Mixology
  • Braun & Clarke, 2022
  • evaluators are always dealing with uncertainty
  • different people have different level of tolerance for uncertainty (and an evaluator’s tolerance might be different than that of the people they work with)
  • aspects of uncertainty”
    • probability – quant representation about the amount of uncertainty
    • ambiguity – different ways to interpret findings
    • vagueness – how detailed the language is
  • take with client before you start – what is the stake of the decision? What is their tolerance for university how certain do they need to be? This can inform choice of methods, etc.
  • uncertainty can be leveraged to drive transformational change by creating dialogue about the unknown and asking more interesting questions about the unknown (e.g., if data is not available, ask why there is no data)

My post conference “to do” list

  • read Decolonizing Wealth
  • read all the various articles that I took note of
  • order System Evaluation Theory: A Blueprint for Practitioners Evaluating Complex Interventions Operating and Functioning as Systems by Ralph Renger (I chatted with Ralph at the book fair, but they didn’t have any books to buy at the book fair – just a chance to talk to authors)

Written by cplysy · Categorized: drbethsnow

May 25 2023

Collective Impact Action Summit Conference 2023

Written by cplysy · Categorized: connectingevidence

May 25 2023

Your Report is a Website

It’s time to stop designing reports for that one person who one time printed and then, subsequently, read your report.

For the vast majority of people, your report is a website. It is read using a digital device, connected to the internet, and served from someone else’s computer (a.k.a. a server probably owned by Amazon).

So what does that mean? Well, it’s time to design your report like the website it truly is, taking advantage of better/more useful technology.

“Did you know that you can create interactive reports in PowerPoint?!? Or PDF! Or Canva! Or Google Docs! Or Tableau! Or Word!”

With linking and basic menus you can make almost anything work like a website.

You can also write a novel with pencil on post it notes. The fact that you CAN do something, doesn’t mean you should.

Creating an effective interactive report using non-web design tools is going to result in a mediocre report design that will likely take much longer than just using a tool like WordPress.

Why WordPress?

  • WordPress powers over half of the internet. It’s open source and has continued to evolve over decades.
  • Using WordPress doesn’t require huge software license investments.
  • There are tons of WordPress developers in the world, so if something goes wrong, you don’t have to find that one expert to fix stuff. Or pay some consultant at Tableau hundreds of dollars an hour to fix what broke.
  • You can do SOOOOO much with a modern WordPress site, even without any plugins. Far more than any of our standard reporting tools.
  • WordPress can easily be built for accessibility.
  • Since WordPress is such a popular platform it has had a profound influence on shaping the evolution of modern web design.
  • You can build both privately and publicly, on in-house servers or through a myriad of cloud based hosting services with the proper encryption.
  • WordPress sites are way better for SEO than any PDF, PowerPoint, or data dashboard tool could hope to match.
  • WordPress sites work really well with social media, email marketing, video, and other modern dissemination channels.

Don’t believe me, I’ll prove it!

This is my new mission (alongside teaching data design).

I’m going to build cutting edge interactive reports and data dashboards using WordPress.

Right now I’m developing out a concept gallery that I can use to show off certain concepts I have only shared with specific clients. Stay tuned.

Written by cplysy · Categorized: freshspectrum

May 24 2023

Letting Go of What No Longer Serves You in Evaluation

When we try to template and checklist our way out of complexity, we lose intentionality and creativity. Wouldn’t you rather build meaningful and sustainable evaluation strategies that ignite your passion and highlight your uniqueness? I know I would. And together we can. This post includes five questions to help you start to reimagine evaluation through intentional inquiry, meaningful measurement, and sustainable strategies.

The post Letting Go of What No Longer Serves You in Evaluation first appeared on Elizabeth Grim Consulting, LLC.

The post Letting Go of What No Longer Serves You in Evaluation appeared first on Elizabeth Grim Consulting, LLC.

Written by cplysy · Categorized: elizabethgrim

May 15 2023

Models of Design Thinking

Design Thinking Mode

Design Thinking is the process we use to conceive of our products and guide the process of design. Design is . This is the design journey. The benefits to engaging with design thinking go beyond the outcomes and include a variety of outputs.

Design thinking is a human-centered approach to innovation that draws from the designer’s toolkit to integrate the needs of people, the possibilities of technology, and the requirements for business success. There are several popular models of design thinking that are widely used in many industries. Here are some of the most common ones:

  1. Stanford d.school’s 5-Stage Model: Stanford d.school is known for its five-stage model of design thinking. The stages are Empathize, Define, Ideate, Prototype, and Test. Each stage is intended to be flexible and iterative, allowing designers to move back and forth between stages as needed.
  2. IDEO’s 3 “I” Model: IDEO, a global design company that popularized design thinking, uses a simplified three-stage model: Inspiration, Ideation, and Implementation. Inspiration is the problem or opportunity that motivates the search for solutions; Ideation involves generating, developing, and testing ideas; and Implementation brings the project to life.
  3. Double Diamond Model by the British Design Council: This model breaks down the design process into four distinct phases – Discover, Define, Develop, and Deliver. The model’s visual representation shows two diamonds, where each diamond represents a problem-solving stage – problem definition and solution generation.
  4. IBM’s Loop Model: IBM’s design thinking model uses a loop system with three main stages: Observe, Reflect, and Make. The stages repeat in a continuous loop, representing the iterative nature of design thinking. IBM also includes principles like focusing on user outcomes, restlessly reinventing, and embracing diversity.
  5. Google Ventures’ Design Sprint: Google Ventures (GV) developed a five-day process for answering critical business questions through design, prototyping, and testing ideas with customers. The five stages are Understand, Sketch, Decide, Prototype, and Validate. This model is often used for fast prototyping and user testing.
  6. The 4D Model by Nigel Cross: Nigel Cross, a notable scholar in design studies, proposed a simplified four-stage model: Discover, Define, Develop, and Deliver. This model is often employed in product and industrial design.
  7. The Design Helix by Cameron Norman and Cense. This is our in-house model designed for living systems and draws on many of the other models. It guides our work and reflects an evolutionary, naturalized, dynamic design process.

Each model has its strengths, and the best one to use can depend on the context and specific needs of the project. It’s also worth noting that while these models provide structure to the design thinking process, design thinking is often more fluid and non-linear than these models suggest.

The Design Helix model assumes that the process gets repeated over time, distinguishing it from other models. It embeds ongoing learning into the model until the purpose or need for the design is completed.

Participation, Co-Creation and Public Engagement

Whatever model you choose, consider that the process of design is often best achieved when it is done with high involvement of the perspective of the intended audience or user. Models like participatory design and co-design are ways to engage those who we are designing things for and are affected by what we create in the design process. This is about designing with, not just for, people.

Below are how the models use collaboration and co-creation:

Collaboration, participation and co-design are integral to most design thinking models but might not always be explicitly outlined. Here’s how collaboration and co-design are represented in each model:

  1. Stanford d.school’s 5-Stage Model: Collaboration is inherent in all stages, especially in the Empathize stage, where designers work closely with users to understand their needs and experiences. Also, the Ideate stage encourages brainstorming with a diverse group of people to generate a wide range of ideas.
  2. IDEO’s 3 “I” Model: Collaboration is embedded in all stages. Designers collaborate with stakeholders and users during the Inspiration stage to gather insights. In the Ideation stage, diverse teams work together to generate and refine ideas. The Implementation stage often involves collaboration between designers, users, and other stakeholders to bring the solution to life.
  3. Double Diamond Model by the British Design Council: Collaboration is crucial in the Discover phase, where designers work with users to understand their needs and experiences. The Develop phase is also highly collaborative, as designers work with each other and stakeholders to refine and improve ideas.
  4. IBM’s Loop Model: The Observe stage involves close collaboration with users to understand their needs. The Reflect stage encourages diverse teams to come together to understand what they’ve observed. The Make stage often involves co-design with users to ensure the solution meets their needs.
  5. Google Ventures’ Design Sprint: The Understand and Sketch stages involve collaboration between the design team and other stakeholders to define the problem and generate potential solutions. The Decide stage involves collaboration to choose the best ideas, and the Validate stage involves testing the solution with users.
  6. The 4D Model by N. Cross: Similar to the others, collaboration is inherent in all stages, particularly in the Discover and Develop phases, where designers work closely with users and other stakeholders.
  7. The Design Helix Model recognizes the benefit of high engagement but does not argue for any specific place for engagement. It recognizes that participation and involvement in the design process is a benefit and often critical to a design, yet it may be difficult or impossible to achieve as desired. Designers must work to determine the strategic and ethical considerations involved in each design process.

In all these models, co-design typically happens during the ideation and prototype stages. Co-design is a process where designers and users (or other stakeholders) work together to create a solution that meets the users’ needs. The idea is that involving stakeholders and prospective ‘users’ in the design process will make the final product more user-friendly and effective.

Design Thinking is a powerful process to channel creativity and facilitate innovation. If you want to implement it in your organization, let’s have a coffee and chat.

The post Models of Design Thinking appeared first on Cense Ltd. .

Written by cplysy · Categorized: cameronnorman

  • « Go to Previous Page
  • Go to page 1
  • Interim pages omitted …
  • Go to page 87
  • Go to page 88
  • Go to page 89
  • Go to page 90
  • Go to page 91
  • Interim pages omitted …
  • Go to page 310
  • Go to Next Page »

Footer

Follow our Work

The easiest way to stay connected to our work is to join our newsletter. You’ll get updates on projects, learn about new events, and hear stories from those evaluators whom the field continues to actively exclude and erase.

Get Updates

Want to take further action or join a pod? Click here to learn more.

Copyright © 2026 · The May 13 Group · Log in

en English
af Afrikaanssq Shqipam አማርኛar العربيةhy Հայերենaz Azərbaycan dilieu Euskarabe Беларуская моваbn বাংলাbs Bosanskibg Българскиca Catalàceb Cebuanony Chichewazh-CN 简体中文zh-TW 繁體中文co Corsuhr Hrvatskics Čeština‎da Dansknl Nederlandsen Englisheo Esperantoet Eestitl Filipinofi Suomifr Françaisfy Fryskgl Galegoka ქართულიde Deutschel Ελληνικάgu ગુજરાતીht Kreyol ayisyenha Harshen Hausahaw Ōlelo Hawaiʻiiw עִבְרִיתhi हिन्दीhmn Hmonghu Magyaris Íslenskaig Igboid Bahasa Indonesiaga Gaeilgeit Italianoja 日本語jw Basa Jawakn ಕನ್ನಡkk Қазақ тіліkm ភាសាខ្មែរko 한국어ku كوردی‎ky Кыргызчаlo ພາສາລາວla Latinlv Latviešu valodalt Lietuvių kalbalb Lëtzebuergeschmk Македонски јазикmg Malagasyms Bahasa Melayuml മലയാളംmt Maltesemi Te Reo Māorimr मराठीmn Монголmy ဗမာစာne नेपालीno Norsk bokmålps پښتوfa فارسیpl Polskipt Portuguêspa ਪੰਜਾਬੀro Românăru Русскийsm Samoangd Gàidhligsr Српски језикst Sesothosn Shonasd سنڌيsi සිංහලsk Slovenčinasl Slovenščinaso Afsoomaalies Españolsu Basa Sundasw Kiswahilisv Svenskatg Тоҷикӣta தமிழ்te తెలుగుth ไทยtr Türkçeuk Українськаur اردوuz O‘zbekchavi Tiếng Việtcy Cymraegxh isiXhosayi יידישyo Yorùbázu Zulu