Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Create dataset loader for Filipino Slang Spelling Normalization #15

Closed
SamuelCahyawijaya opened this issue Nov 1, 2023 · 2 comments · Fixed by #20
Closed

Create dataset loader for Filipino Slang Spelling Normalization #15

SamuelCahyawijaya opened this issue Nov 1, 2023 · 2 comments · Fixed by #20
Assignees

Comments

@SamuelCahyawijaya
Copy link
Collaborator

SamuelCahyawijaya commented Nov 1, 2023

Dataloader name: filipino_slang_norm/filipino_slang_norm.py
DataCatalogue: http://seacrowd.github.io/seacrowd-catalogue/card.html?filipino_slang_norm

Dataset filipino_slang_norm
Description This dataset contains 398 abbreviated and/or contracted Filipino words used in Facebook comments made on weather advisories from a Philippine weather bureau. Each word contains three "correct" versions provided by three undergraduate volunteers.
Subsets -
Languages fil
Tasks Lexical Normalization
License Unknown (unknown)
Homepage https://github.com/ljyflores/efficient-spelling-normalization-filipino
HF URL -
Paper URL https://aclanthology.org/2022.sustainlp-1.5/
@SamuelCahyawijaya SamuelCahyawijaya converted this from a draft issue Nov 1, 2023
@ljvmiranda921
Copy link
Collaborator

Also interested in this!

@ljvmiranda921
Copy link
Collaborator

#self-assign

ljvmiranda921 added a commit to ljvmiranda921/seacrowd-datahub that referenced this issue Nov 2, 2023
SamuelCahyawijaya added a commit that referenced this issue Nov 6, 2023
Closes #15 | Add filipino_slang_norm data loader
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
Status: Done
Development

Successfully merging a pull request may close this issue.

2 participants