LOGO ANALYTICS FOR DECISIONS

11 Tips For Writing a Dissertation Data Analysis

Since the evolution of the fourth industrial revolution – the Digital World; lots of data have surrounded us. There are terabytes of data around us or in data centers that need to be processed and used. The data needs to be appropriately analyzed to process it, and Dissertation data analysis forms its basis. If data analysis is valid and free from errors, the research outcomes will be reliable and lead to a successful dissertation. 

Considering the complexity of many data analysis projects, it becomes challenging to get precise results if analysts are not familiar with data analysis tools and tests properly. The analysis is a time-taking process that starts with collecting valid and relevant data and ends with the demonstration of error-free results.

So, in today’s topic, we will cover the need to analyze data, dissertation data analysis, and mainly the tips for writing an outstanding data analysis dissertation. If you are a doctoral student and plan to perform dissertation data analysis on your data, make sure that you give this article a thorough read for the best tips!

What is Data Analysis in Dissertation?

Dissertation Data Analysis  is the process of understanding, gathering, compiling, and processing a large amount of data. Then identifying common patterns in responses and critically examining facts and figures to find the rationale behind those outcomes.

Even f you have the data collected and compiled in the form of facts and figures, it is not enough for proving your research outcomes. There is still a need to apply dissertation data analysis on your data; to use it in the dissertation. It provides scientific support to the thesis and conclusion of the research.

Data Analysis Tools

There are plenty of indicative tests used to analyze data and infer relevant results for the discussion part. Following are some tests  used to perform analysis of data leading to a scientific conclusion:

11 Most Useful Tips for Dissertation Data Analysis

Doctoral students need to perform dissertation data analysis and then dissertation to receive their degree. Many Ph.D. students find it hard to do dissertation data analysis because they are not trained in it.

1. Dissertation Data Analysis Services

The first tip applies to those students who can afford to look for help with their dissertation data analysis work. It’s a viable option, and it can help with time management and with building the other elements of the dissertation with much detail.

Dissertation Analysis services are professional services that help doctoral students with all the basics of their dissertation work, from planning, research and clarification, methodology, dissertation data analysis and review, literature review, and final powerpoint presentation.

One great reference for dissertation data analysis professional services is Statistics Solutions , they’ve been around for over 22 years helping students succeed in their dissertation work. You can find the link to their website here .

For a proper dissertation data analysis, the student should have a clear understanding and statistical knowledge. Through this knowledge and experience, a student can perform dissertation analysis on their own. 

Following are some helpful tips for writing a splendid dissertation data analysis:

2. Relevance of Collected Data

If the data is irrelevant and not appropriate, you might get distracted from the point of focus. To show the reader that you can critically solve the problem, make sure that you write a theoretical proposition regarding the selection  and analysis of data.

3. Data Analysis

For analysis, it is crucial to use such methods that fit best with the types of data collected and the research objectives. Elaborate on these methods and the ones that justify your data collection methods thoroughly. Make sure to make the reader believe that you did not choose your method randomly. Instead, you arrived at it after critical analysis and prolonged research.

On the other hand,  quantitative analysis  refers to the analysis and interpretation of facts and figures – to build reasoning behind the advent of primary findings. An assessment of the main results and the literature review plays a pivotal role in qualitative and quantitative analysis.

The overall objective of data analysis is to detect patterns and inclinations in data and then present the outcomes implicitly.  It helps in providing a solid foundation for critical conclusions and assisting the researcher to complete the dissertation proposal. 

4. Qualitative Data Analysis

Qualitative data refers to data that does not involve numbers. You are required to carry out an analysis of the data collected through experiments, focus groups, and interviews. This can be a time-taking process because it requires iterative examination and sometimes demanding the application of hermeneutics. Note that using qualitative technique doesn’t only mean generating good outcomes but to unveil more profound knowledge that can be transferrable.

Presenting qualitative data analysis in a dissertation  can also be a challenging task. It contains longer and more detailed responses. Placing such comprehensive data coherently in one chapter of the dissertation can be difficult due to two reasons. Firstly, we cannot figure out clearly which data to include and which one to exclude. Secondly, unlike quantitative data, it becomes problematic to present data in figures and tables. Making information condensed into a visual representation is not possible. As a writer, it is of essence to address both of these challenges.

          Qualitative Data Analysis Methods

Following are the methods used to perform quantitative data analysis. 

  •   Deductive Method

This method involves analyzing qualitative data based on an argument that a researcher already defines. It’s a comparatively easy approach to analyze data. It is suitable for the researcher with a fair idea about the responses they are likely to receive from the questionnaires.

  •  Inductive Method

In this method, the researcher analyzes the data not based on any predefined rules. It is a time-taking process used by students who have very little knowledge of the research phenomenon.

5. Quantitative Data Analysis

Quantitative data contains facts and figures obtained from scientific research and requires extensive statistical analysis. After collection and analysis, you will be able to conclude. Generic outcomes can be accepted beyond the sample by assuming that it is representative – one of the preliminary checkpoints to carry out in your analysis to a larger group. This method is also referred to as the “scientific method”, gaining its roots from natural sciences.

The Presentation of quantitative data  depends on the domain to which it is being presented. It is beneficial to consider your audience while writing your findings. Quantitative data for  hard sciences  might require numeric inputs and statistics. As for  natural sciences , such comprehensive analysis is not required.

                Quantitative Analysis Methods

Following are some of the methods used to perform quantitative data analysis. 

  • Trend analysis:  This corresponds to a statistical analysis approach to look at the trend of quantitative data collected over a considerable period.
  • Cross-tabulation:  This method uses a tabula way to draw readings among data sets in research.  
  • Conjoint analysis :   Quantitative data analysis method that can collect and analyze advanced measures. These measures provide a thorough vision about purchasing decisions and the most importantly, marked parameters.
  • TURF analysis:  This approach assesses the total market reach of a service or product or a mix of both. 
  • Gap analysis:  It utilizes the  side-by-side matrix  to portray quantitative data, which captures the difference between the actual and expected performance. 
  • Text analysis:  In this method, innovative tools enumerate  open-ended data  into easily understandable data. 

6. Data Presentation Tools

Since large volumes of data need to be represented, it becomes a difficult task to present such an amount of data in coherent ways. To resolve this issue, consider all the available choices you have, such as tables, charts, diagrams, and graphs. 

Tables help in presenting both qualitative and quantitative data concisely. While presenting data, always keep your reader in mind. Anything clear to you may not be apparent to your reader. So, constantly rethink whether your data presentation method is understandable to someone less conversant with your research and findings. If the answer is “No”, you may need to rethink your Presentation. 

7. Include Appendix or Addendum

After presenting a large amount of data, your dissertation analysis part might get messy and look disorganized. Also, you would not be cutting down or excluding the data you spent days and months collecting. To avoid this, you should include an appendix part. 

The data you find hard to arrange within the text, include that in the  appendix part of a dissertation . And place questionnaires, copies of focus groups and interviews, and data sheets in the appendix. On the other hand, one must put the statistical analysis and sayings quoted by interviewees within the dissertation. 

8. Thoroughness of Data

It is a common misconception that the data presented is self-explanatory. Most of the students provide the data and quotes and think that it is enough and explaining everything. It is not sufficient. Rather than just quoting everything, you should analyze and identify which data you will use to approve or disapprove your standpoints. 

Thoroughly demonstrate the ideas and critically analyze each perspective taking care of the points where errors can occur. Always make sure to discuss the anomalies and strengths of your data to add credibility to your research.

9. Discussing Data

Discussion of data involves elaborating the dimensions to classify patterns, themes, and trends in presented data. In addition, to balancing, also take theoretical interpretations into account. Discuss the reliability of your data by assessing their effect and significance. Do not hide the anomalies. While using interviews to discuss the data, make sure you use relevant quotes to develop a strong rationale. 

It also involves answering what you are trying to do with the data and how you have structured your findings. Once you have presented the results, the reader will be looking for interpretation. Hence, it is essential to deliver the understanding as soon as you have submitted your data.

10. Findings and Results

Findings refer to the facts derived after the analysis of collected data. These outcomes should be stated; clearly, their statements should tightly support your objective and provide logical reasoning and scientific backing to your point. This part comprises of majority part of the dissertation. 

In the finding part, you should tell the reader what they are looking for. There should be no suspense for the reader as it would divert their attention. State your findings clearly and concisely so that they can get the idea of what is more to come in your dissertation.

11. Connection with Literature Review

At the ending of your data analysis in the dissertation, make sure to compare your data with other published research. In this way, you can identify the points of differences and agreements. Check the consistency of your findings if they meet your expectations—lookup for bottleneck position. Analyze and discuss the reasons behind it. Identify the key themes, gaps, and the relation of your findings with the literature review. In short, you should link your data with your research question, and the questions should form a basis for literature.

The Role of Data Analytics at The Senior Management Level

The Role of Data Analytics at The Senior Management Level

From small and medium-sized businesses to Fortune 500 conglomerates, the success of a modern business is now increasingly tied to how the company implements its data infrastructure and data-based decision-making. According

The Decision-Making Model Explained (In Plain Terms)

The Decision-Making Model Explained (In Plain Terms)

Any form of the systematic decision-making process is better enhanced with data. But making sense of big data or even small data analysis when venturing into a decision-making process might

13 Reasons Why Data Is Important in Decision Making

13 Reasons Why Data Is Important in Decision Making

Wrapping Up

Writing data analysis in the dissertation involves dedication, and its implementations demand sound knowledge and proper planning. Choosing your topic, gathering relevant data, analyzing it, presenting your data and findings correctly, discussing the results, connecting with the literature and conclusions are milestones in it. Among these checkpoints, the Data analysis stage is most important and requires a lot of keenness.

In this article, we thoroughly looked at the tips that prove valuable for writing a data analysis in a dissertation. Make sure to give this article a thorough read before you write data analysis in the dissertation leading to the successful future of your research.

Oxbridge Essays. Top 10 Tips for Writing a Dissertation Data Analysis.

Emidio Amadebai

As an IT Engineer, who is passionate about learning and sharing. I have worked and learned quite a bit from Data Engineers, Data Analysts, Business Analysts, and Key Decision Makers almost for the past 5 years. Interested in learning more about Data Science and How to leverage it for better decision-making in my business and hopefully help you do the same in yours.

Recent Posts

Causal vs Evidential Decision-making (How to Make Businesses More Effective) 

In today’s fast-paced business landscape, it is crucial to make informed decisions to stay in the competition which makes it important to understand the concept of the different characteristics and...

Bootstrapping vs. Boosting

Over the past decade, the field of machine learning has witnessed remarkable advancements in predictive techniques and ensemble learning methods. Ensemble techniques are very popular in machine...

thesis on data analysis

  • Cookies & Privacy
  • GETTING STARTED
  • Introduction
  • FUNDAMENTALS

thesis on data analysis

Getting to the main article

Choosing your route

Setting research questions/ hypotheses

Assessment point

Building the theoretical case

Setting your research strategy

Data collection

Data analysis

Data analysis techniques

In STAGE NINE: Data analysis , we discuss the data you will have collected during STAGE EIGHT: Data collection . However, before you collect your data, having followed the research strategy you set out in this STAGE SIX , it is useful to think about the data analysis techniques you may apply to your data when it is collected.

The statistical tests that are appropriate for your dissertation will depend on (a) the research questions/hypotheses you have set, (b) the research design you are using, and (c) the nature of your data. You should already been clear about your research questions/hypotheses from STAGE THREE: Setting research questions and/or hypotheses , as well as knowing the goal of your research design from STEP TWO: Research design in this STAGE SIX: Setting your research strategy . These two pieces of information - your research questions/hypotheses and research design - will let you know, in principle , the statistical tests that may be appropriate to run on your data in order to answer your research questions.

We highlight the words in principle and may because the most appropriate statistical test to run on your data not only depend on your research questions/hypotheses and research design, but also the nature of your data . As you should have identified in STEP THREE: Research methods , and in the article, Types of variables , in the Fundamentals part of Lærd Dissertation, (a) not all data is the same, and (b) not all variables are measured in the same way (i.e., variables can be dichotomous, ordinal or continuous). In addition, not all data is normal , nor is the data when comparing groups necessarily equal , terms we explain in the Data Analysis section in the Fundamentals part of Lærd Dissertation. As a result, you might think that running a particular statistical test is correct at this point of setting your research strategy (e.g., a statistical test called a dependent t-test ), based on the research questions/hypotheses you have set, but when you collect your data (i.e., during STAGE EIGHT: Data collection ), the data may fail certain assumptions that are important to such a statistical test (i.e., normality and homogeneity of variance ). As a result, you have to run another statistical test (e.g., a Wilcoxon signed-rank test instead of a dependent t-test ).

At this stage in the dissertation process, it is important, or at the very least, useful to think about the data analysis techniques you may apply to your data when it is collected. We suggest that you do this for two reasons:

REASON A Supervisors sometimes expect you to know what statistical analysis you will perform at this stage of the dissertation process

This is not always the case, but if you have had to write a Dissertation Proposal or Ethics Proposal , there is sometimes an expectation that you explain the type of data analysis that you plan to carry out. An understanding of the data analysis that you will carry out on your data can also be an expected component of the Research Strategy chapter of your dissertation write-up (i.e., usually Chapter Three: Research Strategy ). Therefore, it is a good time to think about the data analysis process if you plan to start writing up this chapter at this stage.

REASON B It takes time to get your head around data analysis

When you come to analyse your data in STAGE NINE: Data analysis , you will need to think about (a) selecting the correct statistical tests to perform on your data, (b) running these tests on your data using a statistics package such as SPSS, and (c) learning how to interpret the output from such statistical tests so that you can answer your research questions or hypotheses. Whilst we show you how to do this for a wide range of scenarios in the in the Data Analysis section in the Fundamentals part of Lærd Dissertation, it can be a time consuming process. Unless you took an advanced statistics module/option as part of your degree (i.e., not just an introductory course to statistics, which are often taught in undergraduate and master?s degrees), it can take time to get your head around data analysis. Starting this process at this stage (i.e., STAGE SIX: Research strategy ), rather than waiting until you finish collecting your data (i.e., STAGE EIGHT: Data collection ) is a sensible approach.

Final thoughts...

Setting the research strategy for your dissertation required you to describe, explain and justify the research paradigm, quantitative research design, research method(s), sampling strategy, and approach towards research ethics and data analysis that you plan to follow, as well as determine how you will ensure the research quality of your findings so that you can effectively answer your research questions/hypotheses. However, from a practical perspective, just remember that the main goal of STAGE SIX: Research strategy is to have a clear research strategy that you can implement (i.e., operationalize ). After all, if you are unable to clearly follow your plan and carry out your research in the field, you will struggle to answer your research questions/hypotheses. Once you are sure that you have a clear plan, it is a good idea to take a step back, speak with your supervisor, and assess where you are before moving on to collect data. Therefore, when you are ready, proceed to STAGE SEVEN: Assessment point .

  • Link to facebook
  • Link to linkedin
  • Link to twitter
  • Link to youtube
  • Writing Tips

5 Tips for Handling your Thesis Data Analysis

3-minute read

  • 23rd June 2015

When writing your thesis, the process of analyzing data and working with statistics can be pretty hard at first. This is true whether you’re using specialized data analysis software, like SPSS, or a more descriptive approach. But there are a few guidelines you can follow to make things simpler.

1. Choose the Best Analytical Method for Your Project

The sheer variety of techniques available for data analysis can be confusing! If you are writing a thesis  on internet marketing, for instance, your approach to analysis will be very different to someone writing about biochemistry. As such it is important to adopt an approach appropriate to your research.

2. Double Check Your Methodology

If you are working with quantitative data, it is important to make sure that your analytical techniques are compatible with the methods used to gather your data. Having a clear understanding of what you have done so far will ensure that you achieve accurate results.

For instance, when performing statistical analysis, you may have to choose between parametric and non-parametric testing. If your data is sampled from a population with a broadly Gaussian (i.e., normal) distribution, you will almost always want to use some form of non-parametric testing.

But if you can’t remember or aren’t sure how you selected your sample, you won’t necessarily know the best test to use!

3. Familiarize Yourself with Statistical Analysis and Analytical Software

Thanks to various clever computer programs, you no longer have to be a math genius to conduct top-grade statistical analysis. Nevertheless, learning the basics will help you make informed choices when designing your research and prevent you from making basic mistakes.

Find this useful?

Subscribe to our newsletter and get writing tips from our editors straight to your inbox.

Likewise, trying out different software packages will allow you to pick the one best suited to your needs on your current project.

4. Present Your Data Clearly and Consistently

This is possibly one of the most important parts of writing up your results. Even if your data and statistics are perfect, failure to present your analysis clearly will make it difficult for your reader to follow.

Ask yourself how your analysis would look to someone unfamiliar with your project. If they would be able to understand your analysis, you’re on the right track!

5. Make It Relevant!

Finally, remember that data analysis is about more than just presenting your data. You should also relate your analysis back to your research objectives, discussing its relevance and justifying your interpretations.

This will ensure that your work is easy to follow and demonstrate your understanding of the methods used. So no matter what you are writing about, the analysis is a great time to show off how clever you are!

Share this article:

Post A New Comment

Got content that needs a quick turnaround? Let us polish your work. Explore our editorial business services.

9-minute read

How to Use Infographics to Boost Your Presentation

Is your content getting noticed? Capturing and maintaining an audience’s attention is a challenge when...

8-minute read

Why Interactive PDFs Are Better for Engagement

Are you looking to enhance engagement and captivate your audience through your professional documents? Interactive...

7-minute read

Seven Key Strategies for Voice Search Optimization

Voice search optimization is rapidly shaping the digital landscape, requiring content professionals to adapt their...

4-minute read

Five Creative Ways to Showcase Your Digital Portfolio

Are you a creative freelancer looking to make a lasting impression on potential clients or...

How to Ace Slack Messaging for Contractors and Freelancers

Effective professional communication is an important skill for contractors and freelancers navigating remote work environments....

How to Insert a Text Box in a Google Doc

Google Docs is a powerful collaborative tool, and mastering its features can significantly enhance your...

Logo Harvard University

Make sure your writing is the best it can be with our expert English proofreading and editing.

How do I make a data analysis for my bachelor, master or PhD thesis?

A data analysis is an evaluation of formal data to gain knowledge for the bachelor’s, master’s or doctoral thesis. The aim is to identify patterns in the data, i.e. regularities, irregularities or at least anomalies.

Data can come in many forms, from numbers to the extensive descriptions of objects. As a rule, this data is always in numerical form such as time series or numerical sequences or statistics of all kinds. However, statistics are already processed data.

Data analysis requires some creativity because the solution is usually not obvious. After all, no one has conducted an analysis like this before, or at least you haven't found anything about it in the literature.

The results of a data analysis are answers to initial questions and detailed questions. The answers are numbers and graphics and the interpretation of these numbers and graphics.

What are the advantages of data analysis compared to other methods?

  • Numbers are universal
  • The data is tangible.
  • There are algorithms for calculations and it is easier than a text evaluation.
  • The addressees quickly understand the results.
  • You can really do magic and impress the addressees.
  • It’s easier to visualize the results.

What are the disadvantages of data analysis?

  • Garbage in, garbage out. If the quality of the data is poor, it’s impossible to obtain reliable results.
  • The dependency in data retrieval can be quite annoying. Here are some tips for attracting participants for a survey.
  • You have to know or learn methods or find someone who can help you.
  • Mistakes can be devastating.
  • Missing substance can be detected quickly.
  • Pictures say more than a thousand words. Therefore, if you can’t fill the pages with words, at least throw in graphics. However, usually only the words count.

Under what conditions can or should I conduct a data analysis?

  • If I have to.
  • You must be able to get the right data.
  • If I can perform the calculations myself or at least understand, explain and repeat the calculated evaluations of others.
  • You want a clear personal contribution right from the start.

How do I create the evaluation design for the data analysis?

The most important thing is to ask the right questions, enough questions and also clearly formulated questions. Here are some techniques for asking the right questions:

Good formulation: What is the relationship between Alpha and Beta?

Poor formulation: How are Alpha and Beta related?

Now it’s time for the methods for the calculation. There are dozens of statistical methods, but as always, most calculations can be done with only a handful of statistical methods.

  • Which detailed questions can be formulated as the research question?
  • What data is available? In what format? How is the data prepared?
  • Which key figures allow statements?
  • What methods are available to calculate such indicators? Do my details match? By type (scales), by size (number of records).
  • Do I not need to have a lot of data for a data analysis?

It depends on the media, the questions and the methods I want to use.

A fixed rule is that I need at least 30 data sets for a statistical analysis in order to be able to make representative statements about the population. So statistically it doesn't matter if I have 30 or 30 million records. That's why statistics were invented...

What mistakes do I need to watch out for?

  • Don't do the analysis at the last minute.
  • Formulate questions and hypotheses for evaluation BEFORE data collection!
  • Stay persistent, keep going.
  • Leave the results for a while then revise them.
  • You have to combine theory and the state of research with your results.
  • You must have the time under control

Which tools can I use?

You can use programs of all kinds for calculations. But asking questions is your most powerful aide.

Who can legally help me with a data analysis?

The great intellectual challenge is to develop the research design, to obtain the data and to interpret the results in the end.

Am I allowed to let others perform the calculations?

That's a thing. In the end, every program is useful. If someone else is operating a program, then they can simply be seen as an extension of the program. But this is a comfortable view... Of course, it’s better if you do your own calculations.

A good compromise is to find some help, do a practical calculation then follow the calculation steps meticulously so next time you can do the math yourself. Basically, this functions as a permitted training. One can then justify each step of the calculation in the defense.

What's the best place to start?

Clearly with the detailed questions and hypotheses. These two guide the entire data analysis. So formulate as many detailed questions as possible to answer your main question or research question. You can find detailed instructions and examples for the formulation of these so-called detailed questions in the Thesis Guide.

How does the Aristolo Guide help with data evaluation for the bachelor’s or master’s thesis or dissertation?

The Thesis Guide or Dissertation Guide has instructions for data collection, data preparation, data analysis and interpretation. The guide can also teach you how to formulate questions and answer them with data to create your own experiment. We also have many templates for questionnaires and analyses of all kinds. Good luck writing your text! Silvio and the Aristolo Team PS: Check out the Thesis-ABC and the Thesis Guide for writing a bachelor or master thesis in 31 days.

Thesis-Banner-English-1

  • Deutschland
  • United Kingdom

Dissertation Proofreading Services for a Successful Graduation

  • PhD Dissertations
  • Master’s Dissertations
  • Bachelor’s Dissertations
  • Scientific Dissertations
  • Medical Dissertations
  • Bioscience Dissertations
  • Social Sciences Dissertations
  • Psychology Dissertations
  • Humanities Dissertations
  • Engineering Dissertations
  • Economics Dissertations
  • Service Overview
  • Revisión en inglés
  • Relecture en anglais
  • Revisão em inglês

Manuscript Editing

  • Research Paper Editing
  • Lektorat Doktorarbeit
  • Dissertation Proofreading
  • Englisches Lektorat
  • Journal Manuscript Editing
  • Scientific Manuscript Editing Services
  • Book Manuscript Editing
  • PhD Thesis Proofreading Services
  • Wissenschaftslektorat
  • Korektura anglického textu
  • Akademisches Lektorat
  • Journal Article Editing
  • Manuscript Editing Services

PhD Thesis Editing

  • Medical Editing Sciences
  • Proofreading Rates UK
  • Medical Proofreading
  • PhD Proofreading
  • Academic Proofreading
  • PhD Proofreaders
  • Best Dissertation Proofreaders
  • Masters Dissertation Proofreading
  • Proofreading PhD Thesis Price
  • PhD Dissertation Editing
  • Lektorat Englisch Preise
  • Lektorieren Englisch
  • Wissenschaftliches Lektorat
  • Thesis Proofreading Services
  • PhD Thesis Proofreading
  • Proofreading Thesis Cost
  • Proofreading Thesis
  • Thesis Editing Services
  • Professional Thesis Editing
  • PhD Thesis Editing Services
  • Thesis Editing Cost
  • Dissertation Proofreading Services
  • Proofreading Dissertation

PhD Dissertation Proofreading

  • Dissertation Proofreading Cost
  • Dissertation Proofreader
  • Correção de Artigos Científicos
  • Correção de Trabalhos Academicos
  • Serviços de Correção de Inglês
  • Correção de Dissertação
  • Correção de Textos Precos
  • Revision en Ingles
  • Revision de Textos en Ingles
  • Revision de Tesis
  • Revision Medica en Ingles
  • Revision de Tesis Precio
  • Revisão de Artigos Científicos
  • Revisão de Trabalhos Academicos
  • Serviços de Revisão de Inglês
  • Revisão de Dissertação
  • Revisão de Textos Precos
  • Corrección de Textos en Ingles
  • Corrección de Tesis
  • Corrección de Tesis Precio
  • Corrección Medica en Ingles
  • Corrector ingles
  • Choosing the right Journal
  • Journal Editor’s Feedback
  • Dealing with Rejection
  • Quantitative Research Examples
  • Number of scientific papers published per year
  • Acknowledgements Example
  • ISO, ANSI, CFR & Other
  • Types of Peer Review
  • Withdrawing a Paper
  • What is a good h-index
  • Appendix paper
  • Cover Letter Templates
  • Writing an Article
  • How To Write the Findings
  • Abbreviations: ‘Ibid.’ & ‘Id.’
  • Sample letter to editor for publication
  • Tables and figures in research paper
  • Journal Metrics
  • Revision Process of Journal Publishing
  • JOURNAL GUIDELINES

Select Page

Writing the Data Analysis Chapter(s): Results and Evidence

Posted by Rene Tetzner | Oct 19, 2021 | PhD Success | 0 |

Writing the Data Analysis Chapter(s): Results and Evidence

4.4 Writing the Data Analysis Chapter(s): Results and Evidence

Unlike the introduction, literature review and methodology chapter(s), your results chapter(s) will need to be written for the first time as you draft your thesis even if you submitted a proposal, though this part of your thesis will certainly build upon the preceding chapters. You should have carefully recorded and collected the data (test results, participant responses, computer print outs, observations, transcriptions, notes of various kinds etc.) from your research as you conducted it, so now is the time to review, organise and analyse the data. If your study is quantitative in nature, make sure that you know what all the numbers mean and that you consider them in direct relation to the topic, problem or phenomenon you are investigating, and especially in relation to your research questions and hypotheses. You may find that you require the services of a statistician to help make sense of the data, in which case, obtaining that help sooner rather than later is advisable, because you need to understand your results thoroughly before you can write about them. If, on the other hand, your study is qualitative, you will need to read through the data you have collected several times to become familiar with them both as a whole and in detail so that you can establish important themes, patterns and categories. Remember that ‘qualitative analysis is a creative process and requires thoughtful judgments about what is significant and meaningful in the data’ (Roberts, 2010, p.174; see also Miles & Huberman, 1994) – judgements that often need to be made before the findings can be effectively analysed and presented. If you are combining methodologies in your research, you will also need to consider relationships between the results obtained from the different methods, integrating all the data you have obtained and discovering how the results of one approach support or correlate with the results of another. Ideally, you will have taken careful notes recording your initial thoughts and analyses about the sources you consulted and the results and evidence provided by particular methods and instruments as you put them into practice (as suggested in Sections 2.1.2 and 2.1.4), as these will prove helpful while you consider how best to present your results in your thesis.

Although the ways in which to present and organise the results of doctoral research differ markedly depending on the nature of the study and its findings, as on author and committee preferences and university and department guidelines, there are several basic principles that apply to virtually all theses. First and foremost is the need to present the results of your research both clearly and concisely, and in as objective and factual a manner as possible. There will be time and space to elaborate and interpret your results and speculate on their significance and implications in the final discussion chapter(s) of your thesis, but, generally speaking, such reflection on the meaning of the results should be entirely separate from the factual report of your research findings. There are exceptions, of course, and some candidates, supervisors and departments may prefer the factual presentation and interpretive discussion of results to be blended, just as some thesis topics may demand such treatment, but this is rare and best avoided unless there are persuasive reasons to avoid separating the facts from your thoughts about them. If you do find that you need to blend facts and interpretation in reporting your results, make sure that your language leaves no doubt about the line between the two: words such as ‘seems,’ ‘appears,’ ‘may,’ ‘might,’ probably’ and the like will effectively distinguish analytical speculation from more factual reporting (see also Section 4.5).

thesis on data analysis

You need not dedicate much space in this part of the thesis to the methods you used to arrive at your results because these have already been described in your methodology chapter(s), but they can certainly be revisited briefly to clarify or lend structure to your report. Results are most often presented in a straightforward narrative form which is often supplemented by tables and perhaps by figures such as graphs, charts and maps. An effective approach is to decide immediately which information would be best included in tables and figures, and then to prepare those tables and figures before you begin writing the text for the chapter (see Section 4.4.1 on designing effective tables and figures). Arranging your data into the visually immediate formats provided by tables and figures can, for one, produce interesting surprises by enabling you to see trends and details that you may not have noticed previously, and writing the report of your results will prove easier when you have the tables and figures to work with just as your readers ultimately will. In addition, while the text of the results chapter(s) should certainly highlight the most notable data included in tables and figures, it is essential not to repeat information unnecessarily, so writing with the tables and figures already constructed will help you keep repetition to a minimum. Finally, writing about the tables and figures you create will help you test their clarity and effectiveness for your readers, and you can make any necessary adjustments to the tables and figures as you work. Be sure to refer to each table and figure by number in your text and to make it absolutely clear what you want your readers to see or understand in the table or figure (e.g., ‘see Table 1 for the scores’ and ‘Figure 2 shows this relationship’).

thesis on data analysis

Beyond combining textual narration with the data presented in tables and figures, you will need to organise your report of the results in a manner best suited to the material. You may choose to arrange the presentation of your results chronologically or in a hierarchical order that represents their importance; you might subdivide your results into sections (or separate chapters if there is a great deal of information to accommodate) focussing on the findings of different kinds of methodology (quantitative versus qualitative, for instance) or of different tests, trials, surveys, reviews, case studies and so on; or you may want to create sections (or chapters) focussing on specific themes, patterns or categories or on your research questions and/or hypotheses. The last approach allows you to cluster results that relate to a particular question or hypothesis into a single section and can be particularly useful because it provides cohesion for the thesis as a whole and forces you to focus closely on the issues central to the topic, problem or phenomenon you are investigating. You will, for instance, be able to refer back to the questions and hypotheses presented in your introduction (see Section 3.1), to answer the questions and confirm or dismiss the hypotheses and to anticipate in relation to those questions and hypotheses the discussion and interpretation of your findings that will appear in the next part of the thesis (see Section 4.5). Less effective is an approach that organises the presentation of results according to the items of a survey or questionnaire, because these lend the structure of the instrument used to the results instead of connecting those results directly to the aims, themes and argument of your thesis, but such an organisation can certainly be an important early step in your analysis of the findings and might even be valid for the final thesis if, for instance, your work focuses on developing the instrument involved.

thesis on data analysis

The results generated by doctoral research are unique, and this book cannot hope to outline all the possible approaches for presenting the data and analyses that constitute research results, but it is essential that you devote considerable thought and special care to the way in which you structure the report of your results (Section 6.1 on headings may prove helpful). Whatever structure you choose should accurately reflect the nature of your results and highlight their most important and interesting trends, and it should also effectively allow you (in the next part of the thesis) to discuss and speculate upon your findings in ways that will test the premises of your study, work well in the overall argument of your thesis and lead to significant implications for your research. Regardless of how you organise the main body of your results chapter(s), however, you should include a final paragraph (or more than one paragraph if necessary) that briefly summarises and explains the key results and also guides the reader on to the discussion and interpretation of those results in the following chapter(s).

Why PhD Success?

To Graduate Successfully

This article is part of a book called "PhD Success" which focuses on the writing process of a phd thesis, with its aim being to provide sound practices and principles for reporting and formatting in text the methods, results and discussion of even the most innovative and unique research in ways that are clear, correct, professional and persuasive.

thesis on data analysis

The assumption of the book is that the doctoral candidate reading it is both eager to write and more than capable of doing so, but nonetheless requires information and guidance on exactly what he or she should be writing and how best to approach the task. The basic components of a doctoral thesis are outlined and described, as are the elements of complete and accurate scholarly references, and detailed descriptions of writing practices are clarified through the use of numerous examples.

thesis on data analysis

The basic components of a doctoral thesis are outlined and described, as are the elements of complete and accurate scholarly references, and detailed descriptions of writing practices are clarified through the use of numerous examples. PhD Success provides guidance for students familiar with English and the procedures of English universities, but it also acknowledges that many theses in the English language are now written by candidates whose first language is not English, so it carefully explains the scholarly styles, conventions and standards expected of a successful doctoral thesis in the English language.

thesis on data analysis

Individual chapters of this book address reflective and critical writing early in the thesis process; working successfully with thesis supervisors and benefiting from commentary and criticism; drafting and revising effective thesis chapters and developing an academic or scientific argument; writing and formatting a thesis in clear and correct scholarly English; citing, quoting and documenting sources thoroughly and accurately; and preparing for and excelling in thesis meetings and examinations. 

thesis on data analysis

Completing a doctoral thesis successfully requires long and penetrating thought, intellectual rigour and creativity, original research and sound methods (whether established or innovative), precision in recording detail and a wide-ranging thoroughness, as much perseverance and mental toughness as insight and brilliance, and, no matter how many helpful writing guides are consulted, a great deal of hard work over a significant period of time. Writing a thesis can be an enjoyable as well as a challenging experience, however, and even if it is not always so, the personal and professional rewards of achieving such an enormous goal are considerable, as all doctoral candidates no doubt realise, and will last a great deal longer than any problems that may be encountered during the process.

Interested in Proofreading your PhD Thesis? Get in Touch with us

If you are interested in proofreading your PhD thesis or dissertation, please explore our expert dissertation proofreading services.

thesis on data analysis

Rene Tetzner

Rene Tetzner's blog posts dedicated to academic writing. Although the focus is on How To Write a Doctoral Thesis, many other important aspects of research-based writing, editing and publishing are addressed in helpful detail.

Related Posts

PhD Success – How To Write a Doctoral Thesis

PhD Success – How To Write a Doctoral Thesis

October 1, 2021

Table of Contents – PhD Success

Table of Contents – PhD Success

October 2, 2021

The Essential – Preliminary Matter

The Essential – Preliminary Matter

October 3, 2021

The Main Body of the Thesis

The Main Body of the Thesis

October 4, 2021

CUNY Academic Works

Home > Dissertations, Theses & Capstones Projects by Program > Data Analysis & Visualization Master’s Theses and Capstone Projects

Data Analysis & Visualization Master’s Theses and Capstone Projects

Dissertations/theses/capstones from 2024 2024.

The Charge Forward: An Assessment of Electric Vehicle Charging Infrastructure in New York City , Christopher S. Cali

Visualizing a Life, Uprooted: An Interactive, Web-Map and Scroll-Driven Exploration of the Oral History of my Great-Grandfather – from Ottoman Cilicia to Lebanon and Beyond , Alyssa Campbell

Examining the Health Risks of Particulate Matter 2.5 in New York City: How it Affects Marginalized Groups and the Steps Needed to Reduce Air Pollution , Freddy Castro

Clustering of Patients with Heart Disease , Mukadder Cinar

Modeling of COVID-19 Clinical Outcomes in Mexico: An Analysis of Demographic, Clinical, and Chronic Disease Factors , Livia Clarete

Invisible Hand of Socioeconomic Factors in Rising Trend of Maternal Mortality Rates in the U.S. , Disha Kanada

Multi-Perspective Analysis for Derivative Financial Product Prediction with Stacked Recurrent Neural Networks, Natural Language Processing and Large Language Model , Ethan Lo

What Does One Billion Dollars Look Like?: Visualizing Extreme Wealth , William Mahoney Luckman

Making Sense of Making Parole in New York , Alexandra McGlinchy

Employment Outcomes in Higher Education , Yunxia Wei

Dissertations/Theses/Capstones from 2023 2023

Phantom Shootings , Allan Ambris

Naming Venus: An Exploration of Goddesses, Heroines, and Famous Women , Kavya Beheraj

Social Impacts of Robotics on the Labor and Employment Market , Kelvin Espinal

Fighting the Invisibility of Domestic Violence , Yesenny Fernandez

Navigating Through World’s Military Spending Data with Scroll-Event Driven Visualization , Hong Beom Hur

Evocative Visualization of Void and Fluidity , Tomiko Karino

Analyzing Relationships with Machine Learning , Oscar Ko

Analyzing ‘Fight the Power’ Part 1: Music and Longevity Across Evolving Marketing Eras , Shokolatte Tachikawa

Stand-up Comedy Visualized , Berna Yenidogan

Dissertations/Theses/Capstones from 2022 2022

El Ritmo del Westside: Exploring the Musical Landscape of San Antonio’s Historic Westside , Valeria Alderete

A Comparison of Machine Learning Techniques for Validating Students’ Proficiency in Mathematics , Alexander Avdeev

A Machine Learning Approach to Predicting the Onset of Type II Diabetes in a Sample of Pima Indian Women , Meriem Benarbia

Disrepair, Displacement and Distress: Finding Housing Stories Through Data Visualizations , Jennifer Cheng

Blockchain: Key Principles , Nadezda Chikurova

Data for Power: A Visual Tool for Organizing Unions , Shay Culpepper

Happiness From a Different Perspective , Suparna Das

Happiness and Policy Implications: A Sociological View , Sarah M. Kahl

Heating Fire Incidents in New York City , Merissa K. Lissade

NYC vs. Covid-19: The Human and Financial Resources Deployed to Fight the Most Expensive Health Emergency in History in NYC during the Year 2020 , Elmer A. Maldonado Ramirez

Slices of the Big Apple: A Visual Explanation and Analysis of the New York City Budget , Joanne Ramadani

The Value of NFTs , Angelina Tham

Air Pollution, Climate Change, and Our Health , Kathia Vargas Feliz

Peru's Fishmeal Industry: Its Societal and Environmental Impact , Angel Vizurraga

Why, New York City? Gauging the Quality of Life Through the Thoughts of Tweeters , Sheryl Williams

Dissertations/Theses/Capstones from 2021 2021

Data Analysis and Visualization to Dismantle Gender Discrimination in the Field of Technology , Quinn Bolewicki

Remaking Cinema: Black Hollywood Films, Filmmakers, and Finances , Kiana A. Carrington

Detecting Stance on Covid-19 Vaccine in a Polarized Media , Rodica Ceslov

Dota 2 Hero Selection Analysis , Zhan Gong

An Analysis of Machine Learning Techniques for Economic Recession Prediction , Sheridan Kamal

Black Women in Romance , Vianny C. Lugo Aracena

The Public Innovations Explorer: A Geo-Spatial & Linked-Data Visualization Platform For Publicly Funded Innovation Research In The United States , Seth Schimmel

Making Space for Unquantifiable Data: Hand-drawn Data Visualization , Eva Sibinga

Who Pays? New York State Political Donor Matching with Machine Learning , Annalisa Wilde

  • Colleges, Schools, Centers
  • Disciplines

Advanced Search

  • Notify me via email or RSS

Author Corner

  • Data Analysis & Visualization Program

Home | About | FAQ | My Account | Accessibility Statement

Privacy Copyright

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

  • View all journals
  • My Account Login
  • Explore content
  • About the journal
  • Publish with us
  • Sign up for alerts
  • Data Descriptor
  • Open access
  • Published: 03 May 2024

A dataset for measuring the impact of research data and their curation

  • Libby Hemphill   ORCID: orcid.org/0000-0002-3793-7281 1 , 2 ,
  • Andrea Thomer 3 ,
  • Sara Lafia 1 ,
  • Lizhou Fan 2 ,
  • David Bleckley   ORCID: orcid.org/0000-0001-7715-4348 1 &
  • Elizabeth Moss 1  

Scientific Data volume  11 , Article number:  442 ( 2024 ) Cite this article

595 Accesses

8 Altmetric

Metrics details

  • Research data
  • Social sciences

Science funders, publishers, and data archives make decisions about how to responsibly allocate resources to maximize the reuse potential of research data. This paper introduces a dataset developed to measure the impact of archival and data curation decisions on data reuse. The dataset describes 10,605 social science research datasets, their curation histories, and reuse contexts in 94,755 publications that cover 59 years from 1963 to 2022. The dataset was constructed from study-level metadata, citing publications, and curation records available through the Inter-university Consortium for Political and Social Research (ICPSR) at the University of Michigan. The dataset includes information about study-level attributes (e.g., PIs, funders, subject terms); usage statistics (e.g., downloads, citations); archiving decisions (e.g., curation activities, data transformations); and bibliometric attributes (e.g., journals, authors) for citing publications. This dataset provides information on factors that contribute to long-term data reuse, which can inform the design of effective evidence-based recommendations to support high-impact research data curation decisions.

Similar content being viewed by others

thesis on data analysis

SciSciNet: A large-scale open data lake for the science of science research

thesis on data analysis

Data, measurement and empirical methods in the science of science

thesis on data analysis

Interdisciplinarity revisited: evidence for research impact and dynamism

Background & summary.

Recent policy changes in funding agencies and academic journals have increased data sharing among researchers and between researchers and the public. Data sharing advances science and provides the transparency necessary for evaluating, replicating, and verifying results. However, many data-sharing policies do not explain what constitutes an appropriate dataset for archiving or how to determine the value of datasets to secondary users 1 , 2 , 3 . Questions about how to allocate data-sharing resources efficiently and responsibly have gone unanswered 4 , 5 , 6 . For instance, data-sharing policies recognize that not all data should be curated and preserved, but they do not articulate metrics or guidelines for determining what data are most worthy of investment.

Despite the potential for innovation and advancement that data sharing holds, the best strategies to prioritize datasets for preparation and archiving are often unclear. Some datasets are likely to have more downstream potential than others, and data curation policies and workflows should prioritize high-value data instead of being one-size-fits-all. Though prior research in library and information science has shown that the “analytic potential” of a dataset is key to its reuse value 7 , work is needed to implement conceptual data reuse frameworks 8 , 9 , 10 , 11 , 12 , 13 , 14 . In addition, publishers and data archives need guidance to develop metrics and evaluation strategies to assess the impact of datasets.

Several existing resources have been compiled to study the relationship between the reuse of scholarly products, such as datasets (Table  1 ); however, none of these resources include explicit information on how curation processes are applied to data to increase their value, maximize their accessibility, and ensure their long-term preservation. The CCex (Curation Costs Exchange) provides models of curation services along with cost-related datasets shared by contributors but does not make explicit connections between them or include reuse information 15 . Analyses on platforms such as DataCite 16 have focused on metadata completeness and record usage, but have not included related curation-level information. Analyses of GenBank 17 and FigShare 18 , 19 citation networks do not include curation information. Related studies of Github repository reuse 20 and Softcite software citation 21 reveal significant factors that impact the reuse of secondary research products but do not focus on research data. RD-Switchboard 22 and DSKG 23 are scholarly knowledge graphs linking research data to articles, patents, and grants, but largely omit social science research data and do not include curation-level factors. To our knowledge, other studies of curation work in organizations similar to ICPSR – such as GESIS 24 , Dataverse 25 , and DANS 26 – have not made their underlying data available for analysis.

This paper describes a dataset 27 compiled for the MICA project (Measuring the Impact of Curation Actions) led by investigators at ICPSR, a large social science data archive at the University of Michigan. The dataset was originally developed to study the impacts of data curation and archiving on data reuse. The MICA dataset has supported several previous publications investigating the intensity of data curation actions 28 , the relationship between data curation actions and data reuse 29 , and the structures of research communities in a data citation network 30 . Collectively, these studies help explain the return on various types of curatorial investments. The dataset that we introduce in this paper, which we refer to as the MICA dataset, has the potential to address research questions in the areas of science (e.g., knowledge production), library and information science (e.g., scholarly communication), and data archiving (e.g., reproducible workflows).

We constructed the MICA dataset 27 using records available at ICPSR, a large social science data archive at the University of Michigan. Data set creation involved: collecting and enriching metadata for articles indexed in the ICPSR Bibliography of Data-related Literature against the Dimensions AI bibliometric database; gathering usage statistics for studies from ICPSR’s administrative database; processing data curation work logs from ICPSR’s project tracking platform, Jira; and linking data in social science studies and series to citing analysis papers (Fig.  1 ).

figure 1

Steps to prepare MICA dataset for analysis - external sources are red, primary internal sources are blue, and internal linked sources are green.

Enrich paper metadata

The ICPSR Bibliography of Data-related Literature is a growing database of literature in which data from ICPSR studies have been used. Its creation was funded by the National Science Foundation (Award 9977984), and for the past 20 years it has been supported by ICPSR membership and multiple US federally-funded and foundation-funded topical archives at ICPSR. The Bibliography was originally launched in the year 2000 to aid in data discovery by providing a searchable database linking publications to the study data used in them. The Bibliography collects the universe of output based on the data shared in each study through, which is made available through each ICPSR study’s webpage. The Bibliography contains both peer-reviewed and grey literature, which provides evidence for measuring the impact of research data. For an item to be included in the ICPSR Bibliography, it must contain an analysis of data archived by ICPSR or contain a discussion or critique of the data collection process, study design, or methodology 31 . The Bibliography is manually curated by a team of librarians and information specialists at ICPSR who enter and validate entries. Some publications are supplied to the Bibliography by data depositors, and some citations are submitted to the Bibliography by authors who abide by ICPSR’s terms of use requiring them to submit citations to works in which they analyzed data retrieved from ICPSR. Most of the Bibliography is populated by Bibliography team members, who create custom queries for ICPSR studies performed across numerous sources, including Google Scholar, ProQuest, SSRN, and others. Each record in the Bibliography is one publication that has used one or more ICPSR studies. The version we used was captured on 2021-11-16 and included 94,755 publications.

To expand the coverage of the ICPSR Bibliography, we searched exhaustively for all ICPSR study names, unique numbers assigned to ICPSR studies, and DOIs 32 using a full-text index available through the Dimensions AI database 33 . We accessed Dimensions through a license agreement with the University of Michigan. ICPSR Bibliography librarians and information specialists manually reviewed and validated new entries that matched one or more search criteria. We then used Dimensions to gather enriched metadata and full-text links for items in the Bibliography with DOIs. We matched 43% of the items in the Bibliography to enriched Dimensions metadata including abstracts, field of research codes, concepts, and authors’ institutional information; we also obtained links to full text for 16% of Bibliography items. Based on licensing agreements, we included Dimensions identifiers and links to full text so that users with valid publisher and database access can construct an enriched publication dataset.

Gather study usage data

ICPSR maintains a relational administrative database, DBInfo, that organizes study-level metadata and information on data reuse across separate tables. Studies at ICPSR consist of one or more files collected at a single time or for a single purpose; studies in which the same variables are observed over time are grouped into series. Each study at ICPSR is assigned a DOI, and its metadata are stored in DBInfo. Study metadata follows the Data Documentation Initiative (DDI) Codebook 2.5 standard. DDI elements included in our dataset are title, ICPSR study identification number, DOI, authoring entities, description (abstract), funding agencies, subject terms assigned to the study during curation, and geographic coverage. We also created variables based on DDI elements: total variable count, the presence of survey question text in the metadata, the number of author entities, and whether an author entity was an institution. We gathered metadata for ICPSR’s 10,605 unrestricted public-use studies available as of 2021-11-16 ( https://www.icpsr.umich.edu/web/pages/membership/or/metadata/oai.html ).

To link study usage data with study-level metadata records, we joined study metadata from DBinfo on study usage information, which included total study downloads (data and documentation), individual data file downloads, and cumulative citations from the ICPSR Bibliography. We also gathered descriptive metadata for each study and its variables, which allowed us to summarize and append recoded fields onto the study-level metadata such as curation level, number and type of principle investigators, total variable count, and binary variables indicating whether the study data were made available for online analysis, whether survey question text was made searchable online, and whether the study variables were indexed for search. These characteristics describe aspects of the discoverability of the data to compare with other characteristics of the study. We used the study and series numbers included in the ICPSR Bibliography as unique identifiers to link papers to metadata and analyze the community structure of dataset co-citations in the ICPSR Bibliography 32 .

Process curation work logs

Researchers deposit data at ICPSR for curation and long-term preservation. Between 2016 and 2020, more than 3,000 research studies were deposited with ICPSR. Since 2017, ICPSR has organized curation work into a central unit that provides varied levels of curation that vary in the intensity and complexity of data enhancement that they provide. While the levels of curation are standardized as to effort (level one = less effort, level three = most effort), the specific curatorial actions undertaken for each dataset vary. The specific curation actions are captured in Jira, a work tracking program, which data curators at ICPSR use to collaborate and communicate their progress through tickets. We obtained access to a corpus of 669 completed Jira tickets corresponding to the curation of 566 unique studies between February 2017 and December 2019 28 .

To process the tickets, we focused only on their work log portions, which contained free text descriptions of work that data curators had performed on a deposited study, along with the curators’ identifiers, and timestamps. To protect the confidentiality of the data curators and the processing steps they performed, we collaborated with ICPSR’s curation unit to propose a classification scheme, which we used to train a Naive Bayes classifier and label curation actions in each work log sentence. The eight curation action labels we proposed 28 were: (1) initial review and planning, (2) data transformation, (3) metadata, (4) documentation, (5) quality checks, (6) communication, (7) other, and (8) non-curation work. We note that these categories of curation work are very specific to the curatorial processes and types of data stored at ICPSR, and may not match the curation activities at other repositories. After applying the classifier to the work log sentences, we obtained summary-level curation actions for a subset of all ICPSR studies (5%), along with the total number of hours spent on data curation for each study, and the proportion of time associated with each action during curation.

Data Records

The MICA dataset 27 connects records for each of ICPSR’s archived research studies to the research publications that use them and related curation activities available for a subset of studies (Fig.  2 ). Each of the three tables published in the dataset is available as a study archived at ICPSR. The data tables are distributed as statistical files available for use in SAS, SPSS, Stata, and R as well as delimited and ASCII text files. The dataset is organized around studies and papers as primary entities. The studies table lists ICPSR studies, their metadata attributes, and usage information; the papers table was constructed using the ICPSR Bibliography and Dimensions database; and the curation logs table summarizes the data curation steps performed on a subset of ICPSR studies.

Studies (“ICPSR_STUDIES”): 10,605 social science research datasets available through ICPSR up to 2021-11-16 with variables for ICPSR study number, digital object identifier, study name, series number, series title, authoring entities, full-text description, release date, funding agency, geographic coverage, subject terms, topical archive, curation level, single principal investigator (PI), institutional PI, the total number of PIs, total variables in data files, question text availability, study variable indexing, level of restriction, total unique users downloading study data files and codebooks, total unique users downloading data only, and total unique papers citing data through November 2021. Studies map to the papers and curation logs table through ICPSR study numbers as “STUDY”. However, not every study in this table will have records in the papers and curation logs tables.

Papers (“ICPSR_PAPERS”): 94,755 publications collected from 2000-08-11 to 2021-11-16 in the ICPSR Bibliography and enriched with metadata from the Dimensions database with variables for paper number, identifier, title, authors, publication venue, item type, publication date, input date, ICPSR series numbers used in the paper, ICPSR study numbers used in the paper, the Dimension identifier, and the Dimensions link to the publication’s full text. Papers map to the studies table through ICPSR study numbers in the “STUDY_NUMS” field. Each record represents a single publication, and because a researcher can use multiple datasets when creating a publication, each record may list multiple studies or series.

Curation logs (“ICPSR_CURATION_LOGS”): 649 curation logs for 563 ICPSR studies (although most studies in the subset had one curation log, some studies were associated with multiple logs, with a maximum of 10) curated between February 2017 and December 2019 with variables for study number, action labels assigned to work description sentences using a classifier trained on ICPSR curation logs, hours of work associated with a single log entry, and total hours of work logged for the curation ticket. Curation logs map to the study and paper tables through ICPSR study numbers as “STUDY”. Each record represents a single logged action, and future users may wish to aggregate actions to the study level before joining tables.

figure 2

Entity-relation diagram.

Technical Validation

We report on the reliability of the dataset’s metadata in the following subsections. To support future reuse of the dataset, curation services provided through ICPSR improved data quality by checking for missing values, adding variable labels, and creating a codebook.

All 10,605 studies available through ICPSR have a DOI and a full-text description summarizing what the study is about, the purpose of the study, the main topics covered, and the questions the PIs attempted to answer when they conducted the study. Personal names (i.e., principal investigators) and organizational names (i.e., funding agencies) are standardized against an authority list maintained by ICPSR; geographic names and subject terms are also standardized and hierarchically indexed in the ICPSR Thesaurus 34 . Many of ICPSR’s studies (63%) are in a series and are distributed through the ICPSR General Archive (56%), a non-topical archive that accepts any social or behavioral science data. While study data have been available through ICPSR since 1962, the earliest digital release date recorded for a study was 1984-03-18, when ICPSR’s database was first employed, and the most recent date is 2021-10-28 when the dataset was collected.

Curation level information was recorded starting in 2017 and is available for 1,125 studies (11%); approximately 80% of studies with assigned curation levels received curation services, equally distributed between Levels 1 (least intensive), 2 (moderately intensive), and 3 (most intensive) (Fig.  3 ). Detailed descriptions of ICPSR’s curation levels are available online 35 . Additional metadata are available for a subset of 421 studies (4%), including information about whether the study has a single PI, an institutional PI, the total number of PIs involved, total variables recorded is available for online analysis, has searchable question text, has variables that are indexed for search, contains one or more restricted files, and whether the study is completely restricted. We provided additional metadata for this subset of ICPSR studies because they were released within the past five years and detailed curation and usage information were available for them. Usage statistics including total downloads and data file downloads are available for this subset of studies as well; citation statistics are available for 8,030 studies (76%). Most ICPSR studies have fewer than 500 users, as indicated by total downloads, or citations (Fig.  4 ).

figure 3

ICPSR study curation levels.

figure 4

ICPSR study usage.

A subset of 43,102 publications (45%) available in the ICPSR Bibliography had a DOI. Author metadata were entered as free text, meaning that variations may exist and require additional normalization and pre-processing prior to analysis. While author information is standardized for each publication, individual names may appear in different sort orders (e.g., “Earls, Felton J.” and “Stephen W. Raudenbush”). Most of the items in the ICPSR Bibliography as of 2021-11-16 were journal articles (59%), reports (14%), conference presentations (9%), or theses (8%) (Fig.  5 ). The number of publications collected in the Bibliography has increased each decade since the inception of ICPSR in 1962 (Fig.  6 ). Most ICPSR studies (76%) have one or more citations in a publication.

figure 5

ICPSR Bibliography citation types.

figure 6

ICPSR citations by decade.

Usage Notes

The dataset consists of three tables that can be joined using the “STUDY” key as shown in Fig.  2 . The “ICPSR_PAPERS” table contains one row per paper with one or more cited studies in the “STUDY_NUMS” column. We manipulated and analyzed the tables as CSV files with the Pandas library 36 in Python and the Tidyverse packages 37 in R.

The present MICA dataset can be used independently to study the relationship between curation decisions and data reuse. Evidence of reuse for specific studies is available in several forms: usage information, including downloads and citation counts; and citation contexts within papers that cite data. Analysis may also be performed on the citation network formed between datasets and papers that use them. Finally, curation actions can be associated with properties of studies and usage histories.

This dataset has several limitations of which users should be aware. First, Jira tickets can only be used to represent the intensiveness of curation for activities undertaken since 2017, when ICPSR started using both Curation Levels and Jira. Studies published before 2017 were all curated, but documentation of the extent of that curation was not standardized and therefore could not be included in these analyses. Second, the measure of publications relies upon the authors’ clarity of data citation and the ICPSR Bibliography staff’s ability to discover citations with varying formality and clarity. Thus, there is always a chance that some secondary-data-citing publications have been left out of the bibliography. Finally, there may be some cases in which a paper in the ICSPSR bibliography did not actually obtain data from ICPSR. For example, PIs have often written about or even distributed their data prior to their archival in ICSPR. Therefore, those publications would not have cited ICPSR but they are still collected in the Bibliography as being directly related to the data that were eventually deposited at ICPSR.

In summary, the MICA dataset contains relationships between two main types of entities – papers and studies – which can be mined. The tables in the MICA dataset have supported network analysis (community structure and clique detection) 30 ; natural language processing (NER for dataset reference detection) 32 ; visualizing citation networks (to search for datasets) 38 ; and regression analysis (on curation decisions and data downloads) 29 . The data are currently being used to develop research metrics and recommendation systems for research data. Given that DOIs are provided for ICPSR studies and articles in the ICPSR Bibliography, the MICA dataset can also be used with other bibliometric databases, including DataCite, Crossref, OpenAlex, and related indexes. Subscription-based services, such as Dimensions AI, are also compatible with the MICA dataset. In some cases, these services provide abstracts or full text for papers from which data citation contexts can be extracted for semantic content analysis.

Code availability

The code 27 used to produce the MICA project dataset is available on GitHub at https://github.com/ICPSR/mica-data-descriptor and through Zenodo with the identifier https://doi.org/10.5281/zenodo.8432666 . Data manipulation and pre-processing were performed in Python. Data curation for distribution was performed in SPSS.

He, L. & Han, Z. Do usage counts of scientific data make sense? An investigation of the Dryad repository. Library Hi Tech 35 , 332–342 (2017).

Article   Google Scholar  

Brickley, D., Burgess, M. & Noy, N. Google dataset search: Building a search engine for datasets in an open web ecosystem. In The World Wide Web Conference - WWW ‘19 , 1365–1375 (ACM Press, San Francisco, CA, USA, 2019).

Buneman, P., Dosso, D., Lissandrini, M. & Silvello, G. Data citation and the citation graph. Quantitative Science Studies 2 , 1399–1422 (2022).

Chao, T. C. Disciplinary reach: Investigating the impact of dataset reuse in the earth sciences. Proceedings of the American Society for Information Science and Technology 48 , 1–8 (2011).

Article   ADS   Google Scholar  

Parr, C. et al . A discussion of value metrics for data repositories in earth and environmental sciences. Data Science Journal 18 , 58 (2019).

Eschenfelder, K. R., Shankar, K. & Downey, G. The financial maintenance of social science data archives: Four case studies of long–term infrastructure work. J. Assoc. Inf. Sci. Technol. 73 , 1723–1740 (2022).

Palmer, C. L., Weber, N. M. & Cragin, M. H. The analytic potential of scientific data: Understanding re-use value. Proceedings of the American Society for Information Science and Technology 48 , 1–10 (2011).

Zimmerman, A. S. New knowledge from old data: The role of standards in the sharing and reuse of ecological data. Sci. Technol. Human Values 33 , 631–652 (2008).

Cragin, M. H., Palmer, C. L., Carlson, J. R. & Witt, M. Data sharing, small science and institutional repositories. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences 368 , 4023–4038 (2010).

Article   ADS   CAS   Google Scholar  

Fear, K. M. Measuring and Anticipating the Impact of Data Reuse . Ph.D. thesis, University of Michigan (2013).

Borgman, C. L., Van de Sompel, H., Scharnhorst, A., van den Berg, H. & Treloar, A. Who uses the digital data archive? An exploratory study of DANS. Proceedings of the Association for Information Science and Technology 52 , 1–4 (2015).

Pasquetto, I. V., Borgman, C. L. & Wofford, M. F. Uses and reuses of scientific data: The data creators’ advantage. Harvard Data Science Review 1 (2019).

Gregory, K., Groth, P., Scharnhorst, A. & Wyatt, S. Lost or found? Discovering data needed for research. Harvard Data Science Review (2020).

York, J. Seeking equilibrium in data reuse: A study of knowledge satisficing . Ph.D. thesis, University of Michigan (2022).

Kilbride, W. & Norris, S. Collaborating to clarify the cost of curation. New Review of Information Networking 19 , 44–48 (2014).

Robinson-Garcia, N., Mongeon, P., Jeng, W. & Costas, R. DataCite as a novel bibliometric source: Coverage, strengths and limitations. Journal of Informetrics 11 , 841–854 (2017).

Qin, J., Hemsley, J. & Bratt, S. E. The structural shift and collaboration capacity in GenBank networks: A longitudinal study. Quantitative Science Studies 3 , 174–193 (2022).

Article   PubMed   PubMed Central   Google Scholar  

Acuna, D. E., Yi, Z., Liang, L. & Zhuang, H. Predicting the usage of scientific datasets based on article, author, institution, and journal bibliometrics. In Smits, M. (ed.) Information for a Better World: Shaping the Global Future. iConference 2022 ., 42–52 (Springer International Publishing, Cham, 2022).

Zeng, T., Wu, L., Bratt, S. & Acuna, D. E. Assigning credit to scientific datasets using article citation networks. Journal of Informetrics 14 , 101013 (2020).

Koesten, L., Vougiouklis, P., Simperl, E. & Groth, P. Dataset reuse: Toward translating principles to practice. Patterns 1 , 100136 (2020).

Du, C., Cohoon, J., Lopez, P. & Howison, J. Softcite dataset: A dataset of software mentions in biomedical and economic research publications. J. Assoc. Inf. Sci. Technol. 72 , 870–884 (2021).

Aryani, A. et al . A research graph dataset for connecting research data repositories using RD-Switchboard. Sci Data 5 , 180099 (2018).

Färber, M. & Lamprecht, D. The data set knowledge graph: Creating a linked open data source for data sets. Quantitative Science Studies 2 , 1324–1355 (2021).

Perry, A. & Netscher, S. Measuring the time spent on data curation. Journal of Documentation 78 , 282–304 (2022).

Trisovic, A. et al . Advancing computational reproducibility in the Dataverse data repository platform. In Proceedings of the 3rd International Workshop on Practical Reproducible Evaluation of Computer Systems , P-RECS ‘20, 15–20, https://doi.org/10.1145/3391800.3398173 (Association for Computing Machinery, New York, NY, USA, 2020).

Borgman, C. L., Scharnhorst, A. & Golshan, M. S. Digital data archives as knowledge infrastructures: Mediating data sharing and reuse. Journal of the Association for Information Science and Technology 70 , 888–904, https://doi.org/10.1002/asi.24172 (2019).

Lafia, S. et al . MICA Data Descriptor. Zenodo https://doi.org/10.5281/zenodo.8432666 (2023).

Lafia, S., Thomer, A., Bleckley, D., Akmon, D. & Hemphill, L. Leveraging machine learning to detect data curation activities. In 2021 IEEE 17th International Conference on eScience (eScience) , 149–158, https://doi.org/10.1109/eScience51609.2021.00025 (2021).

Hemphill, L., Pienta, A., Lafia, S., Akmon, D. & Bleckley, D. How do properties of data, their curation, and their funding relate to reuse? J. Assoc. Inf. Sci. Technol. 73 , 1432–44, https://doi.org/10.1002/asi.24646 (2021).

Lafia, S., Fan, L., Thomer, A. & Hemphill, L. Subdivisions and crossroads: Identifying hidden community structures in a data archive’s citation network. Quantitative Science Studies 3 , 694–714, https://doi.org/10.1162/qss_a_00209 (2022).

ICPSR. ICPSR Bibliography of Data-related Literature: Collection Criteria. https://www.icpsr.umich.edu/web/pages/ICPSR/citations/collection-criteria.html (2023).

Lafia, S., Fan, L. & Hemphill, L. A natural language processing pipeline for detecting informal data references in academic literature. Proc. Assoc. Inf. Sci. Technol. 59 , 169–178, https://doi.org/10.1002/pra2.614 (2022).

Hook, D. W., Porter, S. J. & Herzog, C. Dimensions: Building context for search and evaluation. Frontiers in Research Metrics and Analytics 3 , 23, https://doi.org/10.3389/frma.2018.00023 (2018).

https://www.icpsr.umich.edu/web/ICPSR/thesaurus (2002). ICPSR. ICPSR Thesaurus.

https://www.icpsr.umich.edu/files/datamanagement/icpsr-curation-levels.pdf (2020). ICPSR. ICPSR Curation Levels.

McKinney, W. Data Structures for Statistical Computing in Python. In van der Walt, S. & Millman, J. (eds.) Proceedings of the 9th Python in Science Conference , 56–61 (2010).

Wickham, H. et al . Welcome to the Tidyverse. Journal of Open Source Software 4 , 1686 (2019).

Fan, L., Lafia, S., Li, L., Yang, F. & Hemphill, L. DataChat: Prototyping a conversational agent for dataset search and visualization. Proc. Assoc. Inf. Sci. Technol. 60 , 586–591 (2023).

Download references

Acknowledgements

We thank the ICPSR Bibliography staff, the ICPSR Data Curation Unit, and the ICPSR Data Stewardship Committee for their support of this research. This material is based upon work supported by the National Science Foundation under grant 1930645. This project was made possible in part by the Institute of Museum and Library Services LG-37-19-0134-19.

Author information

Authors and affiliations.

Inter-university Consortium for Political and Social Research, University of Michigan, Ann Arbor, MI, 48104, USA

Libby Hemphill, Sara Lafia, David Bleckley & Elizabeth Moss

School of Information, University of Michigan, Ann Arbor, MI, 48104, USA

Libby Hemphill & Lizhou Fan

School of Information, University of Arizona, Tucson, AZ, 85721, USA

Andrea Thomer

You can also search for this author in PubMed   Google Scholar

Contributions

L.H. and A.T. conceptualized the study design, D.B., E.M., and S.L. prepared the data, S.L., L.F., and L.H. analyzed the data, and D.B. validated the data. All authors reviewed and edited the manuscript.

Corresponding author

Correspondence to Libby Hemphill .

Ethics declarations

Competing interests.

The authors declare no competing interests.

Additional information

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ .

Reprints and permissions

About this article

Cite this article.

Hemphill, L., Thomer, A., Lafia, S. et al. A dataset for measuring the impact of research data and their curation. Sci Data 11 , 442 (2024). https://doi.org/10.1038/s41597-024-03303-2

Download citation

Received : 16 November 2023

Accepted : 24 April 2024

Published : 03 May 2024

DOI : https://doi.org/10.1038/s41597-024-03303-2

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

Quick links

  • Explore articles by subject
  • Guide to authors
  • Editorial policies

Sign up for the Nature Briefing newsletter — what matters in science, free to your inbox daily.

thesis on data analysis

thesis on data analysis

Understanding data analysis: A beginner's guide

Before data can be used to tell a story, it must go through a process that makes it usable. Explore the role of data analysis in decision-making.

What is data analysis?

Data analysis is the process of gathering, cleaning, and modeling data to reveal meaningful insights. This data is then crafted into reports that support the strategic decision-making process.

Types of data analysis

There are many different types of data analysis. Each type can be used to answer a different question.

thesis on data analysis

Descriptive analytics

Descriptive analytics refers to the process of analyzing historical data to understand trends and patterns. For example, success or failure to achieve key performance indicators like return on investment.

An example of descriptive analytics is generating reports to provide an overview of an organization's sales and financial data, offering valuable insights into past activities and outcomes.

thesis on data analysis

Predictive analytics

Predictive analytics uses historical data to help predict what might happen in the future, such as identifying past trends in data to determine if they’re likely to recur.

Methods include a range of statistical and machine learning techniques, including neural networks, decision trees, and regression analysis.

thesis on data analysis

Diagnostic analytics

Diagnostic analytics helps answer questions about what caused certain events by looking at performance indicators. Diagnostic analytics techniques supplement basic descriptive analysis.

Generally, diagnostic analytics involves spotting anomalies in data (like an unexpected shift in a metric), gathering data related to these anomalies, and using statistical techniques to identify potential explanations.

thesis on data analysis

Cognitive analytics

Cognitive analytics is a sophisticated form of data analysis that goes beyond traditional methods. This method uses machine learning and natural language processing to understand, reason, and learn from data in a way that resembles human thought processes.

The goal of cognitive analytics is to simulate human-like thinking to provide deeper insights, recognize patterns, and make predictions.

thesis on data analysis

Prescriptive analytics

Prescriptive analytics helps answer questions about what needs to happen next to achieve a certain goal or target. By using insights from prescriptive analytics, organizations can make data-driven decisions in the face of uncertainty.

Data analysts performing prescriptive analysis often rely on machine learning to find patterns in large semantic models and estimate the likelihood of various outcomes.

thesis on data analysis

analyticsText analytics

Text analytics is a way to teach computers to understand human language. It involves using algorithms and other techniques to extract information from large amounts of text data, such as social media posts or customer previews.

Text analytics helps data analysts make sense of what people are saying, find patterns, and gain insights that can be used to make better decisions in fields like business, marketing, and research.

The data analysis process

Compiling and interpreting data so it can be used in decision making is a detailed process and requires a systematic approach. Here are the steps that data analysts follow:

1. Define your objectives.

Clearly define the purpose of your analysis. What specific question are you trying to answer? What problem do you want to solve? Identify your core objectives. This will guide the entire process.

2. Collect and consolidate your data.

Gather your data from all relevant sources using  data analysis software . Ensure that the data is representative and actually covers the variables you want to analyze.

3. Select your analytical methods.

Investigate the various data analysis methods and select the technique that best aligns with your objectives. Many free data analysis software solutions offer built-in algorithms and methods to facilitate this selection process.

4. Clean your data.

Scrutinize your data for errors, missing values, or inconsistencies using the cleansing features already built into your data analysis software. Cleaning the data ensures accuracy and reliability in your analysis and is an important part of data analytics.

5. Uncover valuable insights.

Delve into your data to uncover patterns, trends, and relationships. Use statistical methods, machine learning algorithms, or other analytical techniques that are aligned with your goals. This step transforms raw data into valuable insights.

6. Interpret and visualize the results.

Examine the results of your analyses to understand their implications. Connect these findings with your initial objectives. Then, leverage the visualization tools within free data analysis software to present your insights in a more digestible format.

7. Make an informed decision.

Use the insights gained from your analysis to inform your next steps. Think about how these findings can be utilized to enhance processes, optimize strategies, or improve overall performance.

By following these steps, analysts can systematically approach large sets of data, breaking down the complexities and ensuring the results are actionable for decision makers.

The importance of data analysis

Data analysis is critical because it helps business decision makers make sense of the information they collect in our increasingly data-driven world. Imagine you have a massive pile of puzzle pieces (data), and you want to see the bigger picture (insights). Data analysis is like putting those puzzle pieces together—turning that data into knowledge—to reveal what’s important.

Whether you’re a business decision maker trying to make sense of customer preferences or a scientist studying trends, data analysis is an important tool that helps us understand the world and make informed choices.

Primary data analysis methods

A person working on his desktop an open office environment

Quantitative analysis

Quantitative analysis deals with numbers and measurements (for example, looking at survey results captured through ratings). When performing quantitative analysis, you’ll use mathematical and statistical methods exclusively and answer questions like ‘how much’ or ‘how many.’ 

Two people looking at tablet screen showing a word document

Qualitative analysis

Qualitative analysis is about understanding the subjective meaning behind non-numerical data. For example, analyzing interview responses or looking at pictures to understand emotions. Qualitative analysis looks for patterns, themes, or insights, and is mainly concerned with depth and detail.

Data analysis solutions and resources

Turn your data into actionable insights and visualize the results with ease.

Microsoft 365

Process data and turn ideas into reality with innovative apps, including Excel.

Importance of backing up data

Learn how to back up your data and devices for peace of mind—and added security. 

Copilot in Excel

Go deeper with your data using Microsoft Copilot—your AI assistant.

Excel expense template

Organize and track your business expenses using Excel.

Excel templates

Boost your productivity with free, customizable Excel templates for all types of documents.

Chart designs

Enhance presentations, research, and other materials with customizable chart templates.

Follow Microsoft

 LinkedIn.

Grad Coach

Qualitative Data Analysis Methods 101:

The “big 6” methods + examples.

By: Kerryn Warren (PhD) | Reviewed By: Eunice Rautenbach (D.Tech) | May 2020 (Updated April 2023)

Qualitative data analysis methods. Wow, that’s a mouthful. 

If you’re new to the world of research, qualitative data analysis can look rather intimidating. So much bulky terminology and so many abstract, fluffy concepts. It certainly can be a minefield!

Don’t worry – in this post, we’ll unpack the most popular analysis methods , one at a time, so that you can approach your analysis with confidence and competence – whether that’s for a dissertation, thesis or really any kind of research project.

Qualitative data analysis methods

What (exactly) is qualitative data analysis?

To understand qualitative data analysis, we need to first understand qualitative data – so let’s step back and ask the question, “what exactly is qualitative data?”.

Qualitative data refers to pretty much any data that’s “not numbers” . In other words, it’s not the stuff you measure using a fixed scale or complex equipment, nor do you analyse it using complex statistics or mathematics.

So, if it’s not numbers, what is it?

Words, you guessed? Well… sometimes , yes. Qualitative data can, and often does, take the form of interview transcripts, documents and open-ended survey responses – but it can also involve the interpretation of images and videos. In other words, qualitative isn’t just limited to text-based data.

So, how’s that different from quantitative data, you ask?

Simply put, qualitative research focuses on words, descriptions, concepts or ideas – while quantitative research focuses on numbers and statistics . Qualitative research investigates the “softer side” of things to explore and describe , while quantitative research focuses on the “hard numbers”, to measure differences between variables and the relationships between them. If you’re keen to learn more about the differences between qual and quant, we’ve got a detailed post over here .

qualitative data analysis vs quantitative data analysis

So, qualitative analysis is easier than quantitative, right?

Not quite. In many ways, qualitative data can be challenging and time-consuming to analyse and interpret. At the end of your data collection phase (which itself takes a lot of time), you’ll likely have many pages of text-based data or hours upon hours of audio to work through. You might also have subtle nuances of interactions or discussions that have danced around in your mind, or that you scribbled down in messy field notes. All of this needs to work its way into your analysis.

Making sense of all of this is no small task and you shouldn’t underestimate it. Long story short – qualitative analysis can be a lot of work! Of course, quantitative analysis is no piece of cake either, but it’s important to recognise that qualitative analysis still requires a significant investment in terms of time and effort.

Need a helping hand?

thesis on data analysis

In this post, we’ll explore qualitative data analysis by looking at some of the most common analysis methods we encounter. We’re not going to cover every possible qualitative method and we’re not going to go into heavy detail – we’re just going to give you the big picture. That said, we will of course includes links to loads of extra resources so that you can learn more about whichever analysis method interests you.

Without further delay, let’s get into it.

The “Big 6” Qualitative Analysis Methods 

There are many different types of qualitative data analysis, all of which serve different purposes and have unique strengths and weaknesses . We’ll start by outlining the analysis methods and then we’ll dive into the details for each.

The 6 most popular methods (or at least the ones we see at Grad Coach) are:

  • Content analysis
  • Narrative analysis
  • Discourse analysis
  • Thematic analysis
  • Grounded theory (GT)
  • Interpretive phenomenological analysis (IPA)

Let’s take a look at each of them…

QDA Method #1: Qualitative Content Analysis

Content analysis is possibly the most common and straightforward QDA method. At the simplest level, content analysis is used to evaluate patterns within a piece of content (for example, words, phrases or images) or across multiple pieces of content or sources of communication. For example, a collection of newspaper articles or political speeches.

With content analysis, you could, for instance, identify the frequency with which an idea is shared or spoken about – like the number of times a Kardashian is mentioned on Twitter. Or you could identify patterns of deeper underlying interpretations – for instance, by identifying phrases or words in tourist pamphlets that highlight India as an ancient country.

Because content analysis can be used in such a wide variety of ways, it’s important to go into your analysis with a very specific question and goal, or you’ll get lost in the fog. With content analysis, you’ll group large amounts of text into codes , summarise these into categories, and possibly even tabulate the data to calculate the frequency of certain concepts or variables. Because of this, content analysis provides a small splash of quantitative thinking within a qualitative method.

Naturally, while content analysis is widely useful, it’s not without its drawbacks . One of the main issues with content analysis is that it can be very time-consuming , as it requires lots of reading and re-reading of the texts. Also, because of its multidimensional focus on both qualitative and quantitative aspects, it is sometimes accused of losing important nuances in communication.

Content analysis also tends to concentrate on a very specific timeline and doesn’t take into account what happened before or after that timeline. This isn’t necessarily a bad thing though – just something to be aware of. So, keep these factors in mind if you’re considering content analysis. Every analysis method has its limitations , so don’t be put off by these – just be aware of them ! If you’re interested in learning more about content analysis, the video below provides a good starting point.

QDA Method #2: Narrative Analysis 

As the name suggests, narrative analysis is all about listening to people telling stories and analysing what that means . Since stories serve a functional purpose of helping us make sense of the world, we can gain insights into the ways that people deal with and make sense of reality by analysing their stories and the ways they’re told.

You could, for example, use narrative analysis to explore whether how something is being said is important. For instance, the narrative of a prisoner trying to justify their crime could provide insight into their view of the world and the justice system. Similarly, analysing the ways entrepreneurs talk about the struggles in their careers or cancer patients telling stories of hope could provide powerful insights into their mindsets and perspectives . Simply put, narrative analysis is about paying attention to the stories that people tell – and more importantly, the way they tell them.

Of course, the narrative approach has its weaknesses , too. Sample sizes are generally quite small due to the time-consuming process of capturing narratives. Because of this, along with the multitude of social and lifestyle factors which can influence a subject, narrative analysis can be quite difficult to reproduce in subsequent research. This means that it’s difficult to test the findings of some of this research.

Similarly, researcher bias can have a strong influence on the results here, so you need to be particularly careful about the potential biases you can bring into your analysis when using this method. Nevertheless, narrative analysis is still a very useful qualitative analysis method – just keep these limitations in mind and be careful not to draw broad conclusions . If you’re keen to learn more about narrative analysis, the video below provides a great introduction to this qualitative analysis method.

QDA Method #3: Discourse Analysis 

Discourse is simply a fancy word for written or spoken language or debate . So, discourse analysis is all about analysing language within its social context. In other words, analysing language – such as a conversation, a speech, etc – within the culture and society it takes place. For example, you could analyse how a janitor speaks to a CEO, or how politicians speak about terrorism.

To truly understand these conversations or speeches, the culture and history of those involved in the communication are important factors to consider. For example, a janitor might speak more casually with a CEO in a company that emphasises equality among workers. Similarly, a politician might speak more about terrorism if there was a recent terrorist incident in the country.

So, as you can see, by using discourse analysis, you can identify how culture , history or power dynamics (to name a few) have an effect on the way concepts are spoken about. So, if your research aims and objectives involve understanding culture or power dynamics, discourse analysis can be a powerful method.

Because there are many social influences in terms of how we speak to each other, the potential use of discourse analysis is vast . Of course, this also means it’s important to have a very specific research question (or questions) in mind when analysing your data and looking for patterns and themes, or you might land up going down a winding rabbit hole.

Discourse analysis can also be very time-consuming  as you need to sample the data to the point of saturation – in other words, until no new information and insights emerge. But this is, of course, part of what makes discourse analysis such a powerful technique. So, keep these factors in mind when considering this QDA method. Again, if you’re keen to learn more, the video below presents a good starting point.

QDA Method #4: Thematic Analysis

Thematic analysis looks at patterns of meaning in a data set – for example, a set of interviews or focus group transcripts. But what exactly does that… mean? Well, a thematic analysis takes bodies of data (which are often quite large) and groups them according to similarities – in other words, themes . These themes help us make sense of the content and derive meaning from it.

Let’s take a look at an example.

With thematic analysis, you could analyse 100 online reviews of a popular sushi restaurant to find out what patrons think about the place. By reviewing the data, you would then identify the themes that crop up repeatedly within the data – for example, “fresh ingredients” or “friendly wait staff”.

So, as you can see, thematic analysis can be pretty useful for finding out about people’s experiences , views, and opinions . Therefore, if your research aims and objectives involve understanding people’s experience or view of something, thematic analysis can be a great choice.

Since thematic analysis is a bit of an exploratory process, it’s not unusual for your research questions to develop , or even change as you progress through the analysis. While this is somewhat natural in exploratory research, it can also be seen as a disadvantage as it means that data needs to be re-reviewed each time a research question is adjusted. In other words, thematic analysis can be quite time-consuming – but for a good reason. So, keep this in mind if you choose to use thematic analysis for your project and budget extra time for unexpected adjustments.

Thematic analysis takes bodies of data and groups them according to similarities (themes), which help us make sense of the content.

QDA Method #5: Grounded theory (GT) 

Grounded theory is a powerful qualitative analysis method where the intention is to create a new theory (or theories) using the data at hand, through a series of “ tests ” and “ revisions ”. Strictly speaking, GT is more a research design type than an analysis method, but we’ve included it here as it’s often referred to as a method.

What’s most important with grounded theory is that you go into the analysis with an open mind and let the data speak for itself – rather than dragging existing hypotheses or theories into your analysis. In other words, your analysis must develop from the ground up (hence the name). 

Let’s look at an example of GT in action.

Assume you’re interested in developing a theory about what factors influence students to watch a YouTube video about qualitative analysis. Using Grounded theory , you’d start with this general overarching question about the given population (i.e., graduate students). First, you’d approach a small sample – for example, five graduate students in a department at a university. Ideally, this sample would be reasonably representative of the broader population. You’d interview these students to identify what factors lead them to watch the video.

After analysing the interview data, a general pattern could emerge. For example, you might notice that graduate students are more likely to read a post about qualitative methods if they are just starting on their dissertation journey, or if they have an upcoming test about research methods.

From here, you’ll look for another small sample – for example, five more graduate students in a different department – and see whether this pattern holds true for them. If not, you’ll look for commonalities and adapt your theory accordingly. As this process continues, the theory would develop . As we mentioned earlier, what’s important with grounded theory is that the theory develops from the data – not from some preconceived idea.

So, what are the drawbacks of grounded theory? Well, some argue that there’s a tricky circularity to grounded theory. For it to work, in principle, you should know as little as possible regarding the research question and population, so that you reduce the bias in your interpretation. However, in many circumstances, it’s also thought to be unwise to approach a research question without knowledge of the current literature . In other words, it’s a bit of a “chicken or the egg” situation.

Regardless, grounded theory remains a popular (and powerful) option. Naturally, it’s a very useful method when you’re researching a topic that is completely new or has very little existing research about it, as it allows you to start from scratch and work your way from the ground up .

Grounded theory is used to create a new theory (or theories) by using the data at hand, as opposed to existing theories and frameworks.

QDA Method #6:   Interpretive Phenomenological Analysis (IPA)

Interpretive. Phenomenological. Analysis. IPA . Try saying that three times fast…

Let’s just stick with IPA, okay?

IPA is designed to help you understand the personal experiences of a subject (for example, a person or group of people) concerning a major life event, an experience or a situation . This event or experience is the “phenomenon” that makes up the “P” in IPA. Such phenomena may range from relatively common events – such as motherhood, or being involved in a car accident – to those which are extremely rare – for example, someone’s personal experience in a refugee camp. So, IPA is a great choice if your research involves analysing people’s personal experiences of something that happened to them.

It’s important to remember that IPA is subject – centred . In other words, it’s focused on the experiencer . This means that, while you’ll likely use a coding system to identify commonalities, it’s important not to lose the depth of experience or meaning by trying to reduce everything to codes. Also, keep in mind that since your sample size will generally be very small with IPA, you often won’t be able to draw broad conclusions about the generalisability of your findings. But that’s okay as long as it aligns with your research aims and objectives.

Another thing to be aware of with IPA is personal bias . While researcher bias can creep into all forms of research, self-awareness is critically important with IPA, as it can have a major impact on the results. For example, a researcher who was a victim of a crime himself could insert his own feelings of frustration and anger into the way he interprets the experience of someone who was kidnapped. So, if you’re going to undertake IPA, you need to be very self-aware or you could muddy the analysis.

IPA can help you understand the personal experiences of a person or group concerning a major life event, an experience or a situation.

How to choose the right analysis method

In light of all of the qualitative analysis methods we’ve covered so far, you’re probably asking yourself the question, “ How do I choose the right one? ”

Much like all the other methodological decisions you’ll need to make, selecting the right qualitative analysis method largely depends on your research aims, objectives and questions . In other words, the best tool for the job depends on what you’re trying to build. For example:

  • Perhaps your research aims to analyse the use of words and what they reveal about the intention of the storyteller and the cultural context of the time.
  • Perhaps your research aims to develop an understanding of the unique personal experiences of people that have experienced a certain event, or
  • Perhaps your research aims to develop insight regarding the influence of a certain culture on its members.

As you can probably see, each of these research aims are distinctly different , and therefore different analysis methods would be suitable for each one. For example, narrative analysis would likely be a good option for the first aim, while grounded theory wouldn’t be as relevant. 

It’s also important to remember that each method has its own set of strengths, weaknesses and general limitations. No single analysis method is perfect . So, depending on the nature of your research, it may make sense to adopt more than one method (this is called triangulation ). Keep in mind though that this will of course be quite time-consuming.

As we’ve seen, all of the qualitative analysis methods we’ve discussed make use of coding and theme-generating techniques, but the intent and approach of each analysis method differ quite substantially. So, it’s very important to come into your research with a clear intention before you decide which analysis method (or methods) to use.

Start by reviewing your research aims , objectives and research questions to assess what exactly you’re trying to find out – then select a qualitative analysis method that fits. Never pick a method just because you like it or have experience using it – your analysis method (or methods) must align with your broader research aims and objectives.

No single analysis method is perfect, so it can often make sense to adopt more than one  method (this is called triangulation).

Let’s recap on QDA methods…

In this post, we looked at six popular qualitative data analysis methods:

  • First, we looked at content analysis , a straightforward method that blends a little bit of quant into a primarily qualitative analysis.
  • Then we looked at narrative analysis , which is about analysing how stories are told.
  • Next up was discourse analysis – which is about analysing conversations and interactions.
  • Then we moved on to thematic analysis – which is about identifying themes and patterns.
  • From there, we went south with grounded theory – which is about starting from scratch with a specific question and using the data alone to build a theory in response to that question.
  • And finally, we looked at IPA – which is about understanding people’s unique experiences of a phenomenon.

Of course, these aren’t the only options when it comes to qualitative data analysis, but they’re a great starting point if you’re dipping your toes into qualitative research for the first time.

If you’re still feeling a bit confused, consider our private coaching service , where we hold your hand through the research process to help you develop your best work.

thesis on data analysis

Psst... there’s more!

This post was based on one of our popular Research Bootcamps . If you're working on a research project, you'll definitely want to check this out ...

You Might Also Like:

Research design for qualitative and quantitative studies

84 Comments

Richard N

This has been very helpful. Thank you.

netaji

Thank you madam,

Mariam Jaiyeola

Thank you so much for this information

Nzube

I wonder it so clear for understand and good for me. can I ask additional query?

Lee

Very insightful and useful

Susan Nakaweesi

Good work done with clear explanations. Thank you.

Titilayo

Thanks so much for the write-up, it’s really good.

Hemantha Gunasekara

Thanks madam . It is very important .

Gumathandra

thank you very good

Pramod Bahulekar

This has been very well explained in simple language . It is useful even for a new researcher.

Derek Jansen

Great to hear that. Good luck with your qualitative data analysis, Pramod!

Adam Zahir

This is very useful information. And it was very a clear language structured presentation. Thanks a lot.

Golit,F.

Thank you so much.

Emmanuel

very informative sequential presentation

Shahzada

Precise explanation of method.

Alyssa

Hi, may we use 2 data analysis methods in our qualitative research?

Thanks for your comment. Most commonly, one would use one type of analysis method, but it depends on your research aims and objectives.

Dr. Manju Pandey

You explained it in very simple language, everyone can understand it. Thanks so much.

Phillip

Thank you very much, this is very helpful. It has been explained in a very simple manner that even a layman understands

Anne

Thank nicely explained can I ask is Qualitative content analysis the same as thematic analysis?

Thanks for your comment. No, QCA and thematic are two different types of analysis. This article might help clarify – https://onlinelibrary.wiley.com/doi/10.1111/nhs.12048

Rev. Osadare K . J

This is my first time to come across a well explained data analysis. so helpful.

Tina King

I have thoroughly enjoyed your explanation of the six qualitative analysis methods. This is very helpful. Thank you!

Bromie

Thank you very much, this is well explained and useful

udayangani

i need a citation of your book.

khutsafalo

Thanks a lot , remarkable indeed, enlighting to the best

jas

Hi Derek, What other theories/methods would you recommend when the data is a whole speech?

M

Keep writing useful artikel.

Adane

It is important concept about QDA and also the way to express is easily understandable, so thanks for all.

Carl Benecke

Thank you, this is well explained and very useful.

Ngwisa

Very helpful .Thanks.

Hajra Aman

Hi there! Very well explained. Simple but very useful style of writing. Please provide the citation of the text. warm regards

Hillary Mophethe

The session was very helpful and insightful. Thank you

This was very helpful and insightful. Easy to read and understand

Catherine

As a professional academic writer, this has been so informative and educative. Keep up the good work Grad Coach you are unmatched with quality content for sure.

Keep up the good work Grad Coach you are unmatched with quality content for sure.

Abdulkerim

Its Great and help me the most. A Million Thanks you Dr.

Emanuela

It is a very nice work

Noble Naade

Very insightful. Please, which of this approach could be used for a research that one is trying to elicit students’ misconceptions in a particular concept ?

Karen

This is Amazing and well explained, thanks

amirhossein

great overview

Tebogo

What do we call a research data analysis method that one use to advise or determining the best accounting tool or techniques that should be adopted in a company.

Catherine Shimechero

Informative video, explained in a clear and simple way. Kudos

Van Hmung

Waoo! I have chosen method wrong for my data analysis. But I can revise my work according to this guide. Thank you so much for this helpful lecture.

BRIAN ONYANGO MWAGA

This has been very helpful. It gave me a good view of my research objectives and how to choose the best method. Thematic analysis it is.

Livhuwani Reineth

Very helpful indeed. Thanku so much for the insight.

Storm Erlank

This was incredibly helpful.

Jack Kanas

Very helpful.

catherine

very educative

Wan Roslina

Nicely written especially for novice academic researchers like me! Thank you.

Talash

choosing a right method for a paper is always a hard job for a student, this is a useful information, but it would be more useful personally for me, if the author provide me with a little bit more information about the data analysis techniques in type of explanatory research. Can we use qualitative content analysis technique for explanatory research ? or what is the suitable data analysis method for explanatory research in social studies?

ramesh

that was very helpful for me. because these details are so important to my research. thank you very much

Kumsa Desisa

I learnt a lot. Thank you

Tesfa NT

Relevant and Informative, thanks !

norma

Well-planned and organized, thanks much! 🙂

Dr. Jacob Lubuva

I have reviewed qualitative data analysis in a simplest way possible. The content will highly be useful for developing my book on qualitative data analysis methods. Cheers!

Nyi Nyi Lwin

Clear explanation on qualitative and how about Case study

Ogobuchi Otuu

This was helpful. Thank you

Alicia

This was really of great assistance, it was just the right information needed. Explanation very clear and follow.

Wow, Thanks for making my life easy

C. U

This was helpful thanks .

Dr. Alina Atif

Very helpful…. clear and written in an easily understandable manner. Thank you.

Herb

This was so helpful as it was easy to understand. I’m a new to research thank you so much.

cissy

so educative…. but Ijust want to know which method is coding of the qualitative or tallying done?

Ayo

Thank you for the great content, I have learnt a lot. So helpful

Tesfaye

precise and clear presentation with simple language and thank you for that.

nneheng

very informative content, thank you.

Oscar Kuebutornye

You guys are amazing on YouTube on this platform. Your teachings are great, educative, and informative. kudos!

NG

Brilliant Delivery. You made a complex subject seem so easy. Well done.

Ankit Kumar

Beautifully explained.

Thanks a lot

Kidada Owen-Browne

Is there a video the captures the practical process of coding using automated applications?

Thanks for the comment. We don’t recommend using automated applications for coding, as they are not sufficiently accurate in our experience.

Mathewos Damtew

content analysis can be qualitative research?

Hend

THANK YOU VERY MUCH.

Dev get

Thank you very much for such a wonderful content

Kassahun Aman

do you have any material on Data collection

Prince .S. mpofu

What a powerful explanation of the QDA methods. Thank you.

Kassahun

Great explanation both written and Video. i have been using of it on a day to day working of my thesis project in accounting and finance. Thank you very much for your support.

BORA SAMWELI MATUTULI

very helpful, thank you so much

Submit a Comment Cancel reply

Your email address will not be published. Required fields are marked *

Save my name, email, and website in this browser for the next time I comment.

  • Print Friendly

Purdue University Graduate School

Comparison of Soil Carbon Dynamics Between Restored Prairie and Agricultural Soils in the U.S. Midwest

Globally, soils hold more carbon than both the atmosphere and aboveground terrestrial biosphere combined. Changes in land use and land cover have the potential to alter soil carbon cycling throughout the soil profile, from the surface to meters deep, yet most studies focus only on the near surface impact ( 3 and C 4 photosynthetic pathway plant community composition. Comparative analysis of edaphic properties and soil carbon suggests that deep loess deposits in Nebraska permit enhanced water infiltration and SOC deposition to depths of ~100 cm in 60 years of prairie restoration. In Illinois, poorly drained, clay/lime rich soils on glacial till and a younger restored prairie age (15 years) restricted the influence of prairie restoration to the upper 30 cm. Comparing the δ 13 C values of SOC and SIC in each system demonstrated that SIC at each site is likely of lithogenic origin. This work indicates that the magnitude of influence of restoration management is dependent on edaphic properties inherited from geological and geomorphological controls. Future work should quantify root structures and redox properties to better understand the influence of rooting depth on soil carbon concentrations. Fast-cycling C dynamics can be assessed using continuous, in-situ CO 2 and O 2 soil gas concentration changes. The secondary objective of my thesis was to determine if manual, low temporal resolution gas sampling and analysis are a low cost and effective means of measuring soil O 2 and CO 2 , by comparing it with data from in-situ continuous (hourly) sensors. Manual analysis of soil CO 2 and O 2 from field replicates of buried gas collection cups resulted in measurement differences from the continuous sensors. Measuring CO2 concentration with manual methods often resulted in higher concentrations than hourly, continuous measurements across all sites. Additionally, O 2 concentrations measured by manual methods were higher than hourly values in the restored prairie and less in agricultural sites. A variety of spatial variability, pressure perturbations, calibration offsets, and system leakage influences on both analysis methods could cause the discrepancy.

NSF Grant 1331906

Degree type.

  • Master of Science
  • Earth, Atmospheric and Planetary Sciences

Campus location

  • West Lafayette

Advisor/Supervisor/Committee Chair

Additional committee member 2, additional committee member 3, additional committee member 4, additional committee member 5, usage metrics.

  • Environmental biogeochemistry
  • Soil chemistry and soil carbon sequestration (excl. carbon sequestration science)

CC BY 4.0

IMAGES

  1. sample thesis data analysis and interpretation

    thesis on data analysis

  2. Data analysis section of dissertation. How to Use Quantitative Data Analysis in a Thesis. 2022-10-12

    thesis on data analysis

  3. Data Analytics Thesis Ideas

    thesis on data analysis

  4. sample thesis data analysis and interpretation

    thesis on data analysis

  5. Calaméo

    thesis on data analysis

  6. How to write your dissertation data analysis chapters

    thesis on data analysis

VIDEO

  1. SPSS in Nepali

  2. SPSS in Nepali

  3. SPSS in Nepali

  4. SPSS in Nepali

  5. SPSS Missing Data Analysis of Research Thesis Dissertation in Urdu Hindi

  6. SPSS in Nepali

COMMENTS

  1. 11 Tips For Writing a Dissertation Data Analysis

    And place questionnaires, copies of focus groups and interviews, and data sheets in the appendix. On the other hand, one must put the statistical analysis and sayings quoted by interviewees within the dissertation. 8. Thoroughness of Data. It is a common misconception that the data presented is self-explanatory.

  2. A practical guide to data analysis in general literature reviews

    This article is a practical guide to conducting data analysis in general literature reviews. The general literature review is a synthesis and analysis of published research on a relevant clinical issue, and is a common format for academic theses at the bachelor's and master's levels in nursing, physiotherapy, occupational therapy, public health and other related fields.

  3. Dissertation Results/Findings Chapter (Quantitative)

    The results chapter (also referred to as the findings or analysis chapter) is one of the most important chapters of your dissertation or thesis because it shows the reader what you've found in terms of the quantitative data you've collected. It presents the data using a clear text narrative, supported by tables, graphs and charts.

  4. Dissertation & Thesis Data Analysis Help

    Fast-Track Your Data Analysis, Today. Enter your details below, pop us an email, or book an introductory consultation. If you are a human seeing this field, please leave it empty. Get 1-on-1 help analysing and interpreting your qualitative or quantitative dissertation or thesis data from the experts at Grad Coach. Book online now.

  5. Step 7: Data analysis techniques for your dissertation

    An understanding of the data analysis that you will carry out on your data can also be an expected component of the Research Strategy chapter of your dissertation write-up (i.e., usually Chapter Three: Research Strategy). Therefore, it is a good time to think about the data analysis process if you plan to start writing up this chapter at this ...

  6. How to Write a Results Section

    A results section is where you report the main findings of the data collection and analysis you conducted for your thesis or dissertation. You should report all relevant results concisely and objectively, in a logical order. ... you can structure your results section around key themes or topics that emerged from your analysis of the data. For ...

  7. Quantitative Data Analysis Methods & Techniques 101

    Quantitative data analysis is one of those things that often strikes fear in students. It's totally understandable - quantitative analysis is a complex topic, full of daunting lingo, like medians, modes, correlation and regression.Suddenly we're all wishing we'd paid a little more attention in math class…. The good news is that while quantitative data analysis is a mammoth topic ...

  8. How to Use Quantitative Data Analysis in a Thesis

    This guide discusses the application of quantitative data analysis to your thesis statement. Writing a Strong Thesis Statement. In a relatively short essay of 10 to 15 pages, the thesis statement is generally found in the introductory paragraph. This kind of thesis statement is also typically rather short and straightforward.

  9. 5 Tips for Handling your Thesis Data Analysis

    5 Tips for Handling your Thesis Data Analysis. When writing your thesis, the process of analyzing data and working with statistics can be pretty hard at first. This is true whether you're using specialized data analysis software, like SPSS, or a more descriptive approach. But there are a few guidelines you can follow to make things simpler.

  10. How to write a great data science thesis

    They will stress the importance of structure, substance and style. They will urge you to write down your methodology and results first, then progress to the literature review, introduction and conclusions and to write the summary or abstract last. To write clearly and directly with the reader's expectations always in mind.

  11. How do you analyze qualitative data?

    There are various approaches to qualitative data analysis, but they all share five steps in common: Prepare and organize your data. Review and explore your data. Develop a data coding system. Assign codes to the data. Identify recurring themes. The specifics of each step depend on the focus of the analysis.

  12. A Really Simple Guide to Quantitative Data Analysis

    It is important to know w hat kind of data you are planning to collect or analyse as this w ill. affect your analysis method. A 12 step approach to quantitative data analysis. Step 1: Start with ...

  13. How to make a data analysis in a bachelor, master, PhD thesis?

    A data analysis is an evaluation of formal data to gain knowledge for the bachelor's, master's or doctoral thesis. The aim is to identify patterns in the data, i.e. regularities, irregularities or at least anomalies. Data can come in many forms, from numbers to the extensive descriptions of objects. As a rule, this data is always in ...

  14. The Beginner's Guide to Statistical Analysis

    Table of contents. Step 1: Write your hypotheses and plan your research design. Step 2: Collect data from a sample. Step 3: Summarize your data with descriptive statistics. Step 4: Test hypotheses or make estimates with inferential statistics.

  15. Qualitative Data Analysis Methods for Dissertations

    The method you choose will depend on your research objectives and questions. These are the most common qualitative data analysis methods to help you complete your dissertation: 2. Content analysis: This method is used to analyze documented information from texts, email, media and tangible items.

  16. Writing the Data Analysis Chapter(s): Results and Evidence

    Score 94% Score 94%. 4.4 Writing the Data Analysis Chapter (s): Results and Evidence. Unlike the introduction, literature review and methodology chapter (s), your results chapter (s) will need to be written for the first time as you draft your thesis even if you submitted a proposal, though this part of your thesis will certainly build upon the ...

  17. (PDF) Data analysis: tools and methods

    Good managers can use these tools and data analysis techniques on their own, ... D. Pokorná, Business Data Analyses Possibilities. Diploma thesis. Faculty of Applied Informatics, Tomas Bata ...

  18. PDF Chapter 6: Data Analysis and Interpretation 6.1. Introduction

    interpretation of qualitative data collected for this thesis. 6.2.1 Analysis of qualitative data Qualitative data analysis can be described as the process of making sense from research participants‟ views and opinions of situations, corresponding patterns, themes, categories and ... data analysis well, when he provides the following ...

  19. Data Analysis & Visualization Master's Theses and Capstone Projects

    Data Analysis and Visualization to Dismantle Gender Discrimination in the Field of Technology, Quinn Bolewicki. PDF. Remaking Cinema: Black Hollywood Films, Filmmakers, and Finances, Kiana A. Carrington. PDF. Detecting Stance on Covid-19 Vaccine in a Polarized Media, Rodica Ceslov. PDF. Dota 2 Hero Selection Analysis, Zhan Gong. PDF

  20. PDF Master Thesis: Data Science and Marketing Analytics

    1.3 Structure of the thesis The remainder of the thesis is structured as follows. Chapter 2 gives an overview of the literature focused on attribution modeling in which topics such as attribution modeling, rule-based models, data-driven models, and explainable machine learning are discussed. Chapter 3 briefly discusses

  21. (PDF) CHAPTER FOUR DATA ANALYSIS AND PRESENTATION OF ...

    CHAPTER FOUR. DATA ANALYSIS AND PRESENTATION OF RES EARCH FINDINGS 4.1 Introduction. The chapter contains presentation, analysis and dis cussion of the data collected by the researcher. during the ...

  22. A dataset for measuring the impact of research data and their ...

    Science funders, publishers, and data archives make decisions about how to responsibly allocate resources to maximize the reuse potential of research data. This paper introduces a dataset ...

  23. Research Topics & Ideas: Data Science

    If you're just starting out exploring data science-related topics for your dissertation, thesis or research project, you've come to the right place. In this post, we'll help kickstart your research by providing a hearty list of data science and analytics-related research ideas, including examples from recent studies.. PS - This is just the start…

  24. OMICS Technologies and Data Science in Biomedicine*

    Data analysis in proteomics and mass spectroscopy. Bioinformatic challenges and available software solutions for data analysis in proteomics; ... independently write a master thesis in accordance with content-related and formal academic requirements, reflecting resources and capabilities (time, method comprehension) as well as the access to up ...

  25. Understanding Data Analysis: A Beginner's Guide

    Data analysis is like putting those puzzle pieces together—turning that data into knowledge—to reveal what's important. Whether you're a business decision maker trying to make sense of customer preferences or a scientist studying trends, data analysis is an important tool that helps us understand the world and make informed choices.

  26. Vertiv: Efficiently Absorbing AI Tailwinds (NYSE:VRT)

    Erik Isakson. Investment thesis. Vertiv (NYSE:VRT) is one of the hottest stocks in the U.S. market right now, and its growth was mainly fueled by an AI and data centers mania.When I first saw that ...

  27. Qualitative Data Analysis Methods: Top 6

    QDA Method #3: Discourse Analysis. Discourse is simply a fancy word for written or spoken language or debate. So, discourse analysis is all about analysing language within its social context. In other words, analysing language - such as a conversation, a speech, etc - within the culture and society it takes place.

  28. Master thesis

    I'm doing a Master Thesis, and it's my first time dealing with SEM - PLS 4 (with barely any statistical background behind me). I'm analyzing the relationship between negative word of mouth (NWOM ...

  29. Comparison of Soil Carbon Dynamics Between Restored Prairie and

    The secondary objective of my thesis was to determine if manual, low temporal resolution gas sampling and analysis are a low cost and effective means of measuring soil O2 and CO2, by comparing it with data from in-situ continuous (hourly) sensors.

  30. Commercial, industrial, and institutional discount rate estimation for

    Technical Report: Commercial, industrial, and institutional discount rate estimation for efficiency standards analysis Sector-level data 1998-2023. Commercial, industrial, and ... As part of this determination, DOE conducts a life-cycle cost (LCC) analysis, which models the combined impact of appliance first cost and operating cost changes on ...