university, web hosting, performance (lack thereof)
Wow. One of our uni's research centres maintains a website hosted on Wix that takes 25 seconds to load a single page, and which gets a Lighthouse audit performance score of 5/100.
I understand they prefer the freedom of 3rd party hosting vs. the uni's web constraints, but WTH?
Data analysis software
Trying to convince a colleague who has been blocked for weeks by not having access to SPSS (due to social distancing measures) to try out Python + Pandas or R.
I'm pointing to https://software-carpentry.org/lessons/ as a set of reasonable lessons for someone new to programming.
Hoping they will give it a shot!
While we're all focused on the Corona crisis the US government is quietly pushing an internet #surveillance bill that is aimed at abolishing message encryption:
We should probably pay attention to this and spread awareness even if we are not American: https://www.eff.org/deeplinks/2020/03/earn-it-bill-governments-not-so-secret-plan-scan-every-message-online
Data analysis tools
Having gone through the process of learning pandas (I'm a long-time Pythonista) enough to merge, clean, and recode some survey results, I would feel more comfortable loading the data into PostgreSQL and munging the hell out of it there with CTEs, GROUP BYs, and aggregate functions, etc.
But... having a perfectly replicable, automated process from start to finish is nothing to sneeze at.
Basic data analysis (!)
For the last two weeks our research team has been trying to agree on one number: how many people completed our survey.
I think I've finally convinced them that the number I've given, implemented in Python pandas, is the most correct.
This does not bode well for the real data analysis to follow :/
So just descriptive stats for now. I haven't even started plotting data yet, but I'm looking forward to that.
So much better than fumbling around in a spreadsheet!
I am really liking Python's pandas for cleaning up and analyzing these survey results. I was facing two sets of results for the same survey with responses that were coded differently in each set (due to translation reordering).
So I'm working with the text value of the responses and using https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.DataFrame.replace.html#pandas.DataFrame.replace to get everything into a single language, then doing a first pass of analysis with https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.DataFrame.describe.html#pandas.DataFrame.describe . So far so good! And REPLICABLE!
For context, I discussing this with a research team and was surprised they wanted to keep the partial data. Given the preamble, keeping the data seems unethical.
They have since changed their mind. \o/
Consent preamble on a survey that you're taking said "You can choose to withdraw at any time by closing your browser or navigating away."
After you complete about half of the survey, you decide that you don't want to finish it and close the browser.
What do you expect the researchers to do with the partial data you submitted?
Bilingual survey; data cleanup
I helped design a bilingual survey, but didn't implement it. The PI entered it all into Qualtrics, twice: once for each language.
Now that we have the results, the codes don't match up between languages (of course), so I have to do a ton of cleanup before I can start analysing the data.
But first I have to stop staring at https://www.qualtrics.com/support/survey-platform/survey-module/survey-tools/translate-survey/ wondering why the PI didn't just use that?
See p.8 of https://doi.org/10.5860/lrts.63n2.119 for my current example. (It's paywalled--stupid ALA--but Google Scholar will find a perfect copy for you at academia.edu)
If I zoom in on-screen, I can barely make out the blurry label text.
Worse, even when printed at 600dpi--theoretically the whole reason for the weird layout--fig. 3 is unreadable.
I hate when I have to struggle to read graph labels in PDF articles on even a 27" QHD screen.
There has to be a better way. Like HTML?
Or at least not arbitrarily scaling the graphic down to 5/8s of the page width. Use the full "8.5"" page-width; it's not like an extra page or two in article length is costing you anything.
Today from when you join the review of draft #1 of a collaboratively-edited 30 page whitepaper on how to implement X in an information retrieval system, where X is a cool thing big corporations have been doing for a few years, and ask the question "So has any research been done on whether X actually benefits users? Is there a lit review?", and you get (presumably embarrassed) silence as a response.
Guess who gets to pull together that lit review?
#introduction Hi all :) I'm working on a PhD on #STS, specifically studying the #openhardware for science movement. As an activist I'm interested in feminist approaches to tech which I try to implement in meet ups, workshops, etc. A big part of my time goes to working as open as STS lets me, another big part goes to WikiData because I love it. Beautiful communities are what keep me existing through late capitalism, so here I am in Mastodon <3.
Well, good news is that the data was saved in the database correctly; bad news is that the incorrect data that I was shown & downloaded resulted in my negative evaluation of the tool's reliability & data accuracy. (I finished the survey before the researcher responded to me).
Completed a survey + exercise sent out by a PhD candidate working in the same research space. I found and reported a bug in the exercise that might skew their data significantly.
They're doing really interesting work, and including a hands-on exercise in the middle of the survey was a methodological approach that was new to me. I liked it!
But I hope that bug doesn't screw things up too much--my heart sank when I ran into it.
PhD student (Information Studies) focusing on linked data in library systems. And systems librarian at a university.
The social network of the future: No ads, no corporate surveillance, ethical design, and decentralization! Own your data with Mastodon!