Richa Singh Madnawat Self Assessment

Name: Richa Singh Madnawat

Team: ML Team 8 (Bertinator)

Concise overview of things learned:

It has been a great learning experience even on an initial stage not only on the technical front but also on the functional side of things.

1.Technical Area – Web scraping, Data Mining, Data Preprocessing, Data cleaning, Model Building(Tokenization, Encoding, Padding and Truncating the sentences, Training and Validation Splits, Optimizer & Learning Rate Scheduler, Training)
2.Tools – Scrapy, BeautifulSoup, Python(pandas), Excel, Google Colab
3.Soft Skills – Teamwork, Timeline management, Reporting, Team cohesiveness, Technical communication.

Three achievement highlights:

  1. Working with a Big Data set (53,000+ fields and 13 features).
  2. Scraped data from a website which has an infinite scrolling feature and concatenated all the data frames to set it up for model building.
  3. Understanding of Machine Learning concepts (especially BERT, NLP techniques).

List of meetings/ training attended including social team events:

  1. Weekly Team Meetings (twice per week)
  2. Git Workshop
  3. Python Workshop
  4. Data scraping tutorial within the team
  5. BERT tutorial within the team
  6. Team Building activities within in the usual meetings

Goals for the upcoming week:

  1. Preparing the data for BERT by Tokenizing the sequence
  2. Train the forum classification model by using the DistilBert For Classification model

Detailed statement of tasks done:

  1. Scraped the data from talk.folksy.com forum, including main page content (id, title, created_at, last_posted_at, views, like_count, category_id), post_page (post_id, username, created_at, cooked, post_number, updated_at, reply_count, reply_post_num, reads, topic_id, user_id, topic_slug, forum)
  2. Cleaned the data using Excel, and Python (pandas, regex functions) where it had excess different language, tags, and unwanted data.
  3. Concatenation of data scraped by different teams, performing data cleaning for the next step.
  4. In terms of help, my queries regarding handling infinite scrolling were sorted from the technical as well as project leads of our team by providing me various resources to different tutorials and customized problem-solving solutions.
  5. Imported and fed the datasets to a pandas dataframe, parsed it to ready it for final data cleaning preparation
  6. Filtering null, duplicates, hyperlinks, foreign language in sentences were the tasks included in data cleaning process.
  7. Loading the BERT Tokenizer, Tokenized the sentence, Mapped token to their IDs, padded and truncated the sentences and created attention masks for tokens.
  8. Splitted the training and validation data by randomly selected samples sequentially.
  9. Loaded a pre-trained BERT Model, added a single layer of linear classification on the top, ran the model on PyTorch
  10. Established number of epoch, training steps and created learning rate scheduler.
  11. Ran the training and validation to acheive 96% of accuracy.
  12. Plotted traing and validation loss to access the performance of the model.