: “From learning to earning: 4 essential DataCamp certifications to land your dream job”

: “From learning to earning: 4 essential DataCamp certifications to land your dream job”

From learning to earning: 4 essential DataCamp certifications to land your dream job.

Long-Term Implications and Future Developments of DataCamp Certifications

The increasing importance of data in contemporary business and societal decision-making underscores the growing demand for professionals skilled in data management and analysis. The recently published article “From learning to earning: 4 essential DataCamp certifications to land your dream job” emphasizes the benefit of such qualifications in strengthening career prospects. In this follow-up, we consider the long-term implications and possible future developments of these qualifications.

Long-Term Implications

In the ever-changing digital landscape, a gap between industry requirements and the skillset of the existing workforce often emerges. Having data-specific qualifications like the ones offered in DataCamp can close this gap and offer job-seekers an edge over their counterparts. The increased demand for data literate professionals suggests that these certifications will remain highly valuable.

“The world is growing more data-focused every day, and those with the ability to understand and harness this data will be leading the way.”

The ability to analyze and manage data is indispensable in sectors such as business, healthcare, policy-making and beyond. This high demand across various sectors promises exciting and varied career opportunities for data professionals.

Future Developments

Data professions are continually evolving, and so too will the qualifications required for these roles. Artificial Intelligence and advanced analytics are predicted to reshape the data landscape, and future training programs and certifications will likely need to reflect these changes.

Increased Emphasis on Machine Learning and AI

We foresee a more prominent emphasis on concepts related to Artificial Intelligence (AI) and Machine Learning (ML) in future DataCamp certifications. These fields are projected to take a center stage in the upcoming years, heavily impacting data roles. Professionals equipped with such knowledge will have a competitive advantage.

Development of Ethics and Data Privacy Certifications

We also anticipate a rise in demand for certifications related to ethics and data privacy. With increased public scrutiny towards data management and usage, the need for data professionals with a strong understanding of ethical implications and privacy principles will undoubtedly grow.

Actionable Advice

If you’re planning to fortify your career prospects with data-centric skills, it’s advisable to consider obtaining data certifications, keeping the following pointers in mind:

  1. Stay updated with industry trends: It’s important to remain informed about emerging trends in the data field, ensuring your qualification remains relevant and up-to-date.
  2. Brush up on AI and ML: Given the expected rise of AI and ML in data analysis, getting ahead by familiarizing with these concepts can’t hurt.
  3. Ethics are essential: Be ready to demonstrate your commitment to ethical data use by seeking certificates related to data ethics and privacy.

Taking these steps will ensure that you stay competitive, informed and indispensable in the evolving field of data management.

Read the original article

During a recent discussion with high school juniors and seniors, my goal was to enhance their awareness and understanding of Big (intrusive) Data and AI by focusing on two main areas. The bottom line is that I wanted to turn them into citizens of data science. And to accomplish that, I created a simple exercise… Read More »Creating an AI Utility Function: A Classroom Exercise

Introduction

The intersection between educational structures and emergent technologies, such as Big Data and Artificial Intelligence (AI), has created new approaches towards teaching and learning. A case in focus is the recent initiative to educate high school students on the implications of Big Data and AI, effectively turning them into ‘citizens of data science’. This endeavor includes an AI utility function exercise, designed to foster a deeper understanding of these complex technologies.

Long-term Implications of AI and Big Data in Education

The introduction of AI and Big Data into high school curriculum has significant long-term implications. Enhancing pupils’ awareness and understanding of these concepts is a crucial step towards their inevitable future in a data-driven world. This trend is expected to create global citizens who are not only competent users of data and AI but also understand their potential implications and ethical issues.

Early Exposure to Data Science and AI

One substantial long-term effect of this initiative is that students receive early exposure to data science and AI. This early exposure will likely proliferate the development of technological literacy, enabling the students to navigate and understand a world increasingly shaped by these technologies. It also opens avenues for students who wish to delve deeper into these areas in their future careers.

Solving Real-World Problems

The AI utility function exercise is designed to help students understand how AI works. This will equip them with the potential to apply AI in solving real-world problems in the future.

Possible Future Developments

While the present initiative is a positive step, the potential for its future enhancements and developments is vast.

Progression to More Advanced Topics

As students gain a solid foundation in understanding AI and Big Data, educators may introduce more advanced topics. These can include machine learning, neural networks, and comprehensive data analysis.

Collaboration with Tech Companies

Schools could partner with tech companies for guest lectures, internships, and practical workshops. These companies can offer realistic insights into the AI industry and provide hands-on experience to students.

Actionable Advice

  1. Education policymakers should consider introducing AI and Data Science as standalone subjects, promoting tech literacy from a young age.
  2. Teachers should continually upgrade their knowledge to effectively introduce students to these complex subjects.
  3. Students should actively engage in these learning opportunities to prepare for the technological future.
  4. Tech companies should see this as an opportunity to nurture the next generation of tech talent by partnering with schools and offering hands-on learning experiences.

Creating an AI Utility Function: A Classroom Exercise demonstrates the potential advantage of introducing complex technologies such as AI and Big Data in high school. It underlines the importance of understanding these concepts for the younger generation preparing for a data-driven future.

Read the original article

: Analyzing Preprint Trends in Academic Publishing

: Analyzing Preprint Trends in Academic Publishing

[This article was first published on Rstats – quantixed, and kindly contributed to R-bloggers]. (You can report issue about the content on this page here)


Want to share your content on R-bloggers? click here if you have a blog, or here if you don’t.

Answering the question of what fraction of a journal’s papers were previously available as a preprint is quite difficult to do. The tricky part is matching preprints (from a number of different servers) with the published output from a journal. The easy matches are those that are directly linked together, the remainder though can be hard to identify since the manuscript may change (authors, title, abstract) between the preprint and the published version.

A strategy by Crossref called Marple, that aims match preprints to published outputs seems like the best effort so far. Their code and data up to Aug 2023 is available. Let’s use this to answer the question!

My code is below, let’s look at the results first.

The papers that have a preprint version are red, and those without are in grey. The bars are stacked in these plots and the scale is free so that the journals with different volumes of papers can be compared. The plots show only research papers. Reviews and all other outputs have been excluded as far as possible.

We can replot this to show the fraction of papers that have an associated preprint:

We can see that Elife is on a march to become 100% of papers with preprint version. This is due to a policy decision taken a few years ago.

Then there is a tranche of journals who seem to be stabilising at between 25-50% of outputs having a preprinted version. These journals include: Cell Rep, Dev Cell, Development, EMBO J, J Cell Biol, J Cell Sci, MBoC, Nat Cell Biol, Nat Commun, and Plos Biol.

Finally, journals with a very small fraction of preprinted papers include Cells, FASEB J, Front Cell Dev Biol, JBC.

My focus here was on journals in the cell and developmental biology area. I suspect that the differences in rates between journals reflects the content they carry. Cell and developmental biology, like genetics and biophysics, have an established pattern of preprinting. A journal like JCB, carrying 100% cell biology papers tops out at 50% in 2022. Whereas EMBO J, which has a lower fraction of cell biology papers plateaus at ~30%. However, the discipline doesn’t really explain why Cells and Front Cell Dev Biol have such low preprint rates. I know that there are geographical differences in preprinting and so differences in the regional base of authors at a journal may impact their preprint rate overall. There are likely other contributing factors.

Caveats and things to note:

  • the data only goes up to Aug 2023, so the final bar is unreliable.
  • the assignment is not perfect – there will be some papers here that have a preprint version but are not linked up and some erroneous linkages. I had a sense check of the data for one journal and could see a couple of duplicates in the Crossref data out of ~600 for that journal. So the error rate seems very low.
  • the PubMed data is good but again, it is hard to exclude some outputs that are not research papers if they are not tagged appropriately.

The code

devtools::install_github("ropensci/rentrez")
library(rentrez)
library(XML)

# pre-existing script that parses PubMed XML files
source("Script/pubmedXML.R")

# Fetch papers ----
# search term below exceed 9999 results, so need to use history
srchTrm <- paste('("j cell sci"[ta] OR',
                 '"mol biol cell"[ta] OR',
                 '"j cell biol"[ta] OR',
                 '"nat cell biol"[ta] OR',
                 '"embo j"[ta] OR',
                 '"biochem j"[ta] OR',
                 '"dev cell"[ta] OR',
                 '"faseb j"[ta] OR',
                 '"j biol chem"[ta] OR',
                 '"cells"[ta] OR',
                 '"front cell dev biol"[ta] OR',
                 '"nature communications"[ta] OR',
                 '"cell reports"[ta]) AND',
                 '"development"[ta]) AND',
                 '"elife"[ta]) AND',
                 '"plos biol"[ta]) AND',
                 '(2016 : 2023[pdat]) AND',
                 '(journal article[pt] NOT review[pt])')

# so we will use this
journalSrchTrms <- c('"j cell sci"[ta]','"mol biol cell"[ta]','"j cell biol"[ta]','"nat cell biol"[ta]','"embo j"[ta]',
                     '"biochem j"[ta]','"dev cell"[ta]','"faseb j"[ta]','"j biol chem"[ta]','"cells"[ta]',
                     '"front cell dev biol"[ta]','"nature communications"[ta]','"cell reports"[ta]',
                     '"development"[ta]','"elife"[ta]','"plos biol"[ta]')


# loop through journals and loop through the years
# 2016:2023
pprs <- data.frame()

for (i in 2016:2023) {
  for(j in journalSrchTrms) {
    srchTrm <- paste(j, ' AND ', i, '[pdat]', sep = "")
    pp <- entrez_search(db = "pubmed",
                        term = srchTrm, use_history = TRUE)
    if(pp$count == 0) {
      next
    }
    pp_rec <- entrez_fetch(db = "pubmed", web_history = pp$web_history, rettype = "xml", parsed = TRUE)
    xml_name <- paste("Data/all_", i,"_",extract_jname(j), ".xml", sep = "")
    saveXML(pp_rec, file = xml_name)
    tempdf <- extract_xml_brief(xml_name)
    if(!is.null(tempdf)) {
      pprs <- rbind(pprs, tempdf)
    }
  }
}

Now let’s load in the Crossref data and match it up

library(dplyr)
library(ggplot2)

df_all <- read.csv("Data/crossref-preprint-article-relationships-Aug-2023.csv")

# remove duplicates from pubmed data
pprs <- pprs[!duplicated(pprs$pmid), ]

# remove unwanted publication types by using a vector of strings
unwanted <- c("Review", "Comment", "Retracted Publication", "Retraction of Publication", "Editorial", "Autobiography", "Biography", "Historical", "Published Erratum", "Expression of Concern", "Editorial")
# subset pprs to remove unwanted publication types using grepl
pure <- pprs[!grepl(paste(unwanted, collapse = "|"), pprs$ptype), ]
# ensure that ptype contains "Journal Article"
pure <- pure[grepl("Journal Article", pure$ptype), ]
# remove papers with "NA NA" as the sole author
pure <- pure[!grepl("NA NA", pure$authors), ]

# add factor column to pure that indicates if a row in pprs has a doi that is also found in article_doi
pure$in_crossref <- ifelse(tolower(pure$doi) %in% tolower(df_all$article_doi), "yes", "no")

# find the number of rows in pprs that have a doi that is also found in pure
nrow(pure[pure$in_crossref == "yes",])

# summarize by year the number of papers in pure and how many are in the yes and no category of in_crossref
summary_df <- pure %>%
  # convert from chr to numeric
  mutate(year = as.numeric(year)) %>%
  group_by(year, journal, in_crossref) %>%
  summarise(n = n())

# make a plot to show stacked bars of yes and no for each year
ggplot(summary_df, aes(x = year, y = n, fill = in_crossref)) +
  geom_bar(stat = "identity") +
  theme_minimal() +
  scale_fill_manual(values = c("yes" = "#ae363b", "no" = "#d3d3d3")) +
  lims(x = c(2015.5, 2023.5)) +
  labs(x = "Year", y = "Papers") +
  facet_wrap(~journal, scales = "free_y") +
  theme(legend.position = "none")
ggsave("Output/Plots/preprints_all.png", width = 2400, height = 1800, dpi = 300, units = "px", bg = "white")

# now do plot where the bars stack to 100%
ggplot(summary_df, aes(x = year, y = n, fill = in_crossref)) +
  geom_bar(stat = "identity", position = "fill") +
  theme_minimal() +
  scale_fill_manual(values = c("yes" = "#ae363b", "no" = "#d3d3d3")) +
  lims(x = c(2015.5, 2023.5)) +
  labs(x = "Year", y = "Proportion of papers") +
  facet_wrap(~journal) +
  theme(legend.position = "none")
ggsave("Output/Plots/preprints_scaled.png", width = 2400, height = 1800, dpi = 300, units = "px", bg = "white")

Edit: minor update to first plot and code.

The post title comes from “Pre Self” by Godflesh from the “Post Self” album.

To leave a comment for the author, please follow the link and comment on their blog: Rstats – quantixed.

R-bloggers.com offers daily e-mail updates about R news and tutorials about learning R and many other topics. Click here if you’re looking to post or find an R/data-science job.


Want to share your content on R-bloggers? click here if you have a blog, or here if you don’t.

Continue reading: Pre Self: what fraction of a journal’s papers are preprinted?

Understanding the Growth and Prevalence of Preprinting in Academic Publishing

Preprinting – the sharing of academic papers before peer review – is steadily becoming common practice in many fields. Yet, measuring how many of a journal’s papers are available as preprints can be tricky, due to discrepancies in details like author names, titles, and abstracts that may occur during the transition from preprint to published version. Nonetheless, this task is essential as it allows us to observe the state and trend of preprinting practices, which will significantly influence the future of academia and publishing.

Findings and Analysis

The code discussed in the article attempts to correlate self-archived preprint papers with their corresponding published outputs, using strategies such as Crossref’s Marple. Through analysis, we observe that some journals, like Elife, are almost entirely composed of papers with preprint versions due to policy shifts favoring preprinting within the past few years. A second group of publications, including widely recognized journals such as Cell Reports, Development, and Nat Commun, have around 25-50% of their content originating from preprints. However, several journals, like Cells and FASEB J, show very low preprint rates.

The rate differences between journals could be influenced by the specificity of their subjects. Fields like cell and developmental biology – which are quite established in their preprinting practices – tend to feature higher rates of preprint originality. For instance, the Journal of Cell Biology’s (JCB) preprint rates reached 50% in 2022, while EMBO Journal – a journal with lesser focus on cell biology – only reaches around 30%. Geographic differences among the author base, alongside other undefined factors, could also affect preprint rates.

Long-term Implications and Future Developments

The rise of preprinting… practices presents several potential implications for the academic and publishing communities. If current trends persist or accelerate, we could see a more open and transparent academic landscape where the sharing of pioneering research does not have to wait for the lengthy publishing process. However, it also raises issues, such as the credibility of non-peer-reviewed papers and potential ‘scooping’ of research ideas.

At this rate, the publishing world may need to revise its policies and practices to accommodate and properly manage these changes. Incorporating more robust measures for preprint and published article matching could help improve data analytics and reporting in academic publishing. Furthermore, efforts to standardize preprinting practices may help alleviate some concerns or issues born out of its rapid adoption.

Actionable Advice

Scholars and researchers are advised to stay updated on preprinting practices in their respective fields. Preprinting can provide more immediate visibility to your research, but careful consideration should be given to the potential drawbacks. Further, journals and publishers should reassess their approaches to preprints, taking steps to more accurately account for the shift to this new publishing model.

Lastly, developers, data analysts, and librarians could cross-reference this code with their data to extract meaningful insights about preprint practices in their respective fields or institutions. This data will help keep these stakeholders informed and facilitate more strategic decision-making processes in line with the changing nature of academic publishing.

Read the original article

: “Understanding Regularization in Machine Learning and Deep Learning”

: “Understanding Regularization in Machine Learning and Deep Learning”

This article explains the concept of regularization and its significance in machine learning and deep learning. We have discussed how regularization can be used to enhance the performance of linear models, as well as how it can be applied to improve the performance of deep learning models.

Understanding the Impact of Regularization on Machine Learning and Deep Learning

Our exploration into the world of machine learning and deep learning brings up one significant aspect – regularization. This process plays a key role in improving the performance of both linear models and deep learning models. Through this article, we will delve deeper into the future implications of regularization and how it may evolve as these technologies continue to expand.

Long-Term Implications of Regularization

Given the current trends and advances in machine learning and deep learning algorithms, regularization is poised to have lasting implications. One major aspect is in the prevention of overfitting, where models are too closely fit to the training data, hence underperforming on unseen data. As machine learning and deep learning become more sophisticated, regularization will become increasingly vital in mitigating this issue.

Furthermore, with the exponential growth of data, models are becoming increasingly complex, fuelling overfitting and increasing computational costs. Through regularization, we can better manage this complexity, keeping computational requirements in check and enhancing the overall performance of the models.

Possible Future Developments

As technology and data science evolve, so too does our need for improved regularization techniques. We may soon witness the development of new regularization algorithms that are more effective in preventing overfitting and reducing computational costs.

Notably, there may be advancements in adaptive regularization techniques that adjust the regularization strength based on the complexity of the model. This would lead to smarter and more efficient machine learning and deep learning models.

Actionable Advice

  1. Stay Informed: As regularization becomes increasingly important, it’s crucial to stay updated on the latest research and methodologies. New techniques could bring significant benefits to your machine learning and deep learning models.
  2. Invest in Training: Regularization is a nuanced aspect of machine learning and deep learning. It’s vital to understand it fully to leverage its benefits. Continuous training and learning opportunities for your team will ensure you’re well-equipped.
  3. Adopt Best Practices: Implement regularization in your workflows to enhance model performance. It’s a preventive measure against overfitting and a method for improving prediction accuracy on unseen data.
  4. Future-proof Your Machine Learning Models: Invest in developing or adopting adaptive regularization techniques that can adjust to the complexity of the model. This will make your models both effective and efficient.

Read the original article

As we approach 2024, no brand or customer remains unaffected by AI. Our interactions in the marketplace have altered as a result.

Understanding the Impact of AI by 2024

Whether directly or indirectly, it is undeniable that artificial intelligence has touched virtually every aspect of our lives, including the business sector. As we approach 2024, no brand or customer is unaffected by AI. This is leading to a significant transformation in the marketplace dynamics, altering our interactions with businesses and other consumers.

Long-Term Implications of AI in Businesses

The Dawn of Personalized Shopping

AI-backed personalization is poised to change the shopping experience as we know it. With the capacity to process and analyze vast customer data, AI can accurately predict buying behavior and preferences, creating a more personalized shopping experience. This essentially means businesses will win customers over based not just on product quality, but also on the ability to provide an enriched, tailored experience.

Improved Customer Service

Artificial intelligence is shaping the future of customer service. Chatbots integrated with AI capabilities can effectively respond to customer queries 24/7, which not only improves customer satisfaction, but also brings significant cost savings for businesses. This consistent AI-efficiency promises to refine the future of the customer service industry.

Augmenting Decision Making

AI technology is increasingly intersecting with decision-making processes in businesses. By providing predictive analysis and valuable insights, AI can guide businesses to make informed decisions, reducing risks and maximizing profitability.

Possible Future Developments

Even though the impact of AI on business is already profound, it is bound to become more prominent in the future. We may foresee advancements in AI technology such as self-improving algorithms, an increase in the adoption of AI in small and medium enterprises (SMEs), and a perpetual evolution in customer service practices.

Actionable Advice

Given this trend, there are several strategies that businesses can adopt to remain competitive:

  1. Invest in AI Technology: Businesses need to understand that AI is not just a fad, but an investment with substantial return potential. Enterprises should consider integrating AI within their business strategy and improve their operational efficiency.
  2. Data Management: Handling and processing data is a key part of AI integrations. To implement AI solutions effectively, businesses need to have robust data management systems in place. This would allow them to fully leverage AI’s power for predictive analysis and decision making.
  3. Focus on Personalization: With AI’s ability to create tailored experiences, businesses should make personalization a core aspect of their strategy. This not only enhances customer satisfaction but also boosts brand loyalty.
  4. Continual Learning: The field of AI is constantly evolving. Businesses need to stay abreast with the latest developments and trends in AI technology to maintain a competitive edge.

As we embark upon the future, we must acknowledge and respect the role of artificial intelligence in reshaping our world. For businesses, the task at hand is clear: to adapt, evolve, and thrive in this new AI-dominated landscape.

Read the original article

: Cleveland R User Group: Navigating Pandemic Adaptations and Baseball Analytics

: Cleveland R User Group: Navigating Pandemic Adaptations and Baseball Analytics

[This article was first published on R Consortium, and kindly contributed to R-bloggers]. (You can report issue about the content on this page here)


Want to share your content on R-bloggers? click here if you have a blog, or here if you don’t.

Last year, R Consortium talked to John Blischak and Tim Hoolihan of the Cleveland R User Group about their regular structured and casual virtual meetups during the pandemic. Recently, Alec Wong, another co-organizer of the Cleveland R User Group, updated the R Consortium about how the group provides a networking platform for a small but vibrant local R community. Alec shared details of a recent event from the group regarding the use of R for analyzing baseball data. He also discussed some tools for keeping the group inclusive and improving communication among group members.

Please share about your background and involvement with the RUGS group.

I completed my Bachelor of Science degree in Fisheries and Wildlife from the University of Nebraska-Lincoln in 2013, and my Master of Science degree in Statistical Ecology from Cornell University in late 2018. During my graduate program, I gained extensive experience using R, which is the de facto language of the ecological sciences. I discovered a passion for the language, as it is extremely intuitive and pleasant to work with.

After completing my program in 2018, I moved to Cleveland and immediately began attending the Cleveland R User Group in 2019, and have been a consistent member ever since. I eagerly look forward to each of our events. 

After completing my graduate program, I started working at Progressive Insurance. Working for a large organization like Progressive provides me with many diverse opportunities to make use of my extensive experience with R. I was happy to find a vibrant R community within the company, which allowed me to connect with other R users, share knowledge, and I enthusiastically offer one-on-one assistance to analysts from all over Progressive.

Starting in 2022, I accepted the role of co-organizer of the Cleveland R User Group. As a co-organizer, I help with various tasks related to organizing events, such as the one we held last September. I am passionate about fostering the growth of these communities and helping to attract more individuals who enjoy using R.

Our group events are currently being held in a hybrid format. When we manage to find space, we will meet in person, such as when we met to view the 2023 posit::conf in October–several members visited in person and watched and discussed videos from the conference. Most of our meetups continue to be virtual, including our Saturday morning coffee meetups, but we are actively searching for a more permanent physical space to accommodate our regular meetups. 

I am only one of several co-organizers of the Cleveland R user group. The other co-organizers include Tim Hoolihan from Centric Consulting, John Blischak who operates his consulting firm JDB Software Consulting, LLC, and Jim Hester, currently a Senior Software Engineer at Netflix. Their contributions are invaluable and the community benefits tremendously from their efforts.

Can you share what the R community is like in Cleveland? 

I believe interest in R has been fairly steady over time in Cleveland since 2019. We have a handful of members who attend regularly, and typically each meeting one or two new attendees will introduce themselves. 

I would venture to say that R continues to be used frequently in academic settings in Cleveland, though I am ‌unfamiliar with the standards at local universities. At least two of our members belong to local universities and they use R in their curricula. 

As for industry usage, many local companies, including Progressive use R. At Progressive, we have a small, but solid R community; although it is not as large as the Python community, I believe that the R community is more vibrant. This seems characteristic of R communities in varying contexts, as far as I’ve seen. Another Cleveland company, the Cleveland Guardians baseball team, makes use of R for data science. In September 2023 we were fortunate to invite one of their principal data scientists to speak to us about their methods and analyses. (More details below.)

Typically, our attendance is local to the greater Cleveland area, but with virtual meetups, we’ve been able to host speakers and attendees from across the country; this was a silver lining of the pandemic. We also hold regular Saturday morning coffee and informal chat sessions, and it’s great to see fresh faces from outside Cleveland joining in.

You had a Meetup titled “How Major League Teams Use R to Analyze Baseball Data”, can you share more on the topic covered? Why this topic?

On September 27th, 2023, we invited Keith Woolner, principal data scientist at the Cleveland Guardians baseball team, to give a presentation to our group. This was our first in-person meetup after the pandemic, and Progressive generously sponsored our event, affording us a large presentation space, food, and A/V support. We entertained a mixed audience from the public as well as Progressive employees.

Keith spoke to us about “How Major League Baseball Teams Use R to Analyze Baseball Data.” In an engaging session, he showcased several statistical methods used in sports analytics, the code used to produce these analyses, and visualizations of the data and statistical methods. Of particular interest to me was his analysis using a generalized additive model (GAM) to evaluate the relative performance of catchers’ ability to “frame” a catch; in other words, their ability to convince the umpire a strike occurred. The presentation held some relevance for everyone, whether they were interested in Cleveland baseball, statistics, or R, making it a terrific option for our first in-person presentation since January 2020. His presentation drove a lot of engagement both during and after the session.

Any techniques you recommend using for planning for or during the event? (Github, zoom, other) Can these techniques be used to make your group more inclusive to people that are unable to attend physical events in the future?  

One of our co-organizers, John Blischak, has created a slick website using GitHub Pages to showcase our group and used GitHub issue templates to create a process for speakers to submit talks. Additionally, the Cleveland R User group has posted recordings of our meetups to YouTube since 2017, increasing our visibility and accessibility. Many people at Progressive could not attend our September meetup and asked for the recording of our September 2023 meetup as soon as it was available.

Recently, we have also created a Discord server, a platform similar to Slack. This was suggested by one of our members, Ken Wong, and it has been a great addition to our community. We have been growing the server organically since October of last year by marketing it to attendees who visit our events, particularly on the Saturday morning meetups. This has opened up an additional space for us to collaborate and share content asynchronously. Ken has done an excellent job of organizing the server and has added some automated processes that post from R blogs, journal articles, and tweets from high-profile R users. Overall, we are pleased with our progress and look forward to continuing to improve our initiatives.

How do I Join?

R Consortium’s R User Group and Small Conference Support Program (RUGS) provides grants to help R groups organize, share information, and support each other worldwide. We have given grants over the past four years, encompassing over 68,000 members in 33 countries. We would like to include you! Cash grants and meetup.com accounts are awarded based on the intended use of the funds and the amount of money available to distribute.

The post The Cleveland R User Group’s Journey Through Pandemic Adaptations and Baseball Analytics appeared first on R Consortium.

To leave a comment for the author, please follow the link and comment on their blog: R Consortium.

R-bloggers.com offers daily e-mail updates about R news and tutorials about learning R and many other topics. Click here if you’re looking to post or find an R/data-science job.


Want to share your content on R-bloggers? click here if you have a blog, or here if you don’t.

Continue reading: The Cleveland R User Group’s Journey Through Pandemic Adaptations and Baseball Analytics

Cleveland R User Group: Embracing Hybrid Models and R Analytics in Baseball

The Cleveland R User Group, co-organized by Alec Wong, has been actively navigating the shifting dynamics of community involvement during the pandemic, with regular virtual meetups and post-pandemic hybrid models. A recently spotlighted event discussed the use of R for analyzing baseball data. This article explores the key details of the event, the use of R in both academic and industrial settings within Cleveland, and how the group is heightening inclusivity and communication methods.

Use of R in Cleveland

According to Wong, the usage and interest of R in Cleveland has remained steady since 2019. While it’s particularly prevalent in academic environments, the programming language is also utilized by several companies, including Progressive Insurance where Wong works. Additionally, the Cleveland Guardians baseball team uses R for data science applications.

Local and Remote Involvement

The Cleveland R User Group regularly holds meetups in hybrid format. While some members prefer to meet in person, the majority of the meetings take place virtually. The user group is actively searching for a permanent physical meeting space. This virtual trend paved the way to host speakers and attendees from across the country, extending the reach outside of Cleveland.

Event Spotlight: Using R to Analyze Baseball Data

The group recently held an event on September 27th, 2023, titled “How Major League Teams Use R to Analyze Baseball Data,” with Keith Woolner, the principal data scientist at the Cleveland Guardians baseball team. Keith illustrated several statistical methods used in sports analytics with R, including the use of a generalized additive model to evaluate the performance of catchers’ ability.

Greater Inclusivity and Improved Communication

The Cleveland R User Group is working on enhancing inclusivity and improving communication among its members by leveraging technologies like GitHub and Discord. John Blischak, a fellow co-organizer of the team, has developed a website using GitHub Pages, and the team has been posting recordings of their meetups on YouTube to improve accessibility. Recently, a Discord server was created to provide a platform for collaboration and content sharing among community members.

Actionable Advice

  1. Encourage Hybrid Meetups: Companies and communities alike shouldn’t hesitate to continue embracing virtual platforms for increased accessibility and wider reach even post-pandemic.
  2. Utilize Digital Tools for Inclusivity: By leveraging digital platforms like GitHub and Discord, communities like the Cleveland R User Group can streamline communications, improve visibility, and promote inclusivity.
  3. Apply for Grants: For similar user groups or communities, it might be worth scrambling to the relevance of R Consortium’s R User Group and Small Conference Support Program (RUGS) that offers grants to help R groups organize.
  4. Exploit the Power of R: With versatile use cases of R in different industries, it’s an opportunity for academia and businesses to keep exploring and harnessing the power of R for both simple and complex analytical tasks.

Read the original article