IQSS logo

IRC log for #dataverse, 2020-06-15

Connect via chat.dataverse.org to discuss Dataverse (dataverse.org, an open source web application for sharing, citing, analyzing, and preserving research data) with users and developers.

| Channels | #dataverse index | Today | | Search | Google Search | Plain-Text | plain, newest first | summary

All times shown according to UTC.

Time S Nick Message
05:50 jri joined #dataverse
06:31 jri joined #dataverse
07:03 jri joined #dataverse
07:33 jri_ joined #dataverse
07:43 Tommy13 joined #dataverse
08:46 Tommy13 Hello! As I see, Dataverse exports and sends only a few metadata to DataCite. If I want to use more fields, I have to modify the code as the currently used fields are hardcoded. I think it would be better to make the export more generic, so we could use the dataset fields as placeholders in the datacite_metadata_template. Is there any work in progress on DataCite export? What is its state?
09:18 poikilotherm1 Hi Tommy13
09:19 poikilotherm1 Welcome to the Dataverse Community
09:20 poikilotherm1 The only thing I'm aware of regarding DataCite is the support for draft DOIs being on its way to main
09:20 poikilotherm1 (or at least develop)
09:21 poikilotherm1 Looks like there are a few issues open regarding the DataCite support
09:22 poikilotherm1 Oh and I dunno if using the old non-REST API from DataCite might implicate some restricitions on the metadata side of things.
11:15 yoh joined #dataverse
11:18 juancorr joined #dataverse
11:28 stefankasberger joined #dataverse
11:32 donsizemore joined #dataverse
11:57 stefankasberger hi guys. long time no see. :)
12:04 donsizemore @stefankasberger welcome!
12:42 poikilotherm1 Good to have you around again :-)
12:57 poikilotherm1 Hi stefankasberger
13:48 andrewSC joined #dataverse
14:55 pdurbin joined #dataverse
15:02 pdurbin Those hard-coded fields drive me crazy. We should be much more dynamic with regard to the fields we send to DataCite.
15:15 pdurbin donsizemore poikilotherm1 et al.: How do we feel about installing Dataverse on Windows? Impossible? Easy? https://groups.google.com/d/msg/dataverse-community/IKtgm4p-6j8/p-dmlFU6BQAJ
15:16 * donsizemore touches nose
15:16 pdurbin :)
15:26 poikilotherm1 pdurbin they might try with WSL2
15:27 pdurbin Does WSL2 run on servers? Or is it just a desktop thing?
15:28 * pdurbin heads to the design meeting
15:31 poikilotherm1 pdurbin containers might be another option on Windoze
15:34 poikilotherm1 It should run on servers. But I'm not sure if this is production ready
16:25 poikilotherm1 pdurbin I just moved https://github.com/IQSS/dataverse/pull/6974 from draft to code review :-)
16:33 pdurbin poikilotherm1: this confuses me: "Please be aware that non-verified mail addresses will exclude the user even if matching." Do you mean "Please be aware that non-verified mail addresses are excluded."?
16:56 poikilotherm1 Maybe. I just wanted to make it crystal clear that the user excluded, not the mail address
16:56 poikilotherm1 Once it's verified, it will be included
17:03 jri joined #dataverse
17:53 pdurbin Thanks, that helps.
18:00 donsizemore @pdurbin bzzzt, got a sec?
18:01 pdurbin yup
18:01 donsizemore the extant dataverse-metrics site pulls from miniverse, which is returning no JSON
18:02 donsizemore should i point it at https://github.com/IQSS/dataverse-metrics/blob/master/all-dataverse-installations.json instead?
18:02 pdurbin Let me make sure I updated it.
18:03 pdurbin Yes. And the idea is that this script updates that JSON file: https://github.com/IQSS/dataverse-metrics/blob/master/update-all-installations-list.sh
18:04 pdurbin Did I tell you a new IQSS-wide metrics site is coming? More than just Dataverse. Other projects, etc.
18:05 pdurbin stefankasberger: still there?
18:07 donsizemore @pdurbin you didn't, but i'm all ears
18:08 donsizemore @pdurbin in our cron job, currently update-all-installations-list.sh comes after metrics.py, which does the updating. ima switch those.
18:08 pdurbin Well, the site isn't up yet but it's somewhat based on the Dataverse one. Maybe we should pull you into a meeting with the guy building it sometime. So you can give feedback on if this design works or not.
18:09 pdurbin donsizemore: sure makes sense to switch 'em.
18:36 donsizemore @pdurbin java 14 was arbitrary, merely because it's current. i'm happy to switch that to 17
18:37 pdurbin cool, I'm about to ask in #dv-tech
18:38 pdurbin when we're done talking about the backlog :)
18:54 donsizemore i'm on board with 8,11,17
19:04 nightowl313 joined #dataverse
19:06 nightowl313 hi all ... wondering if i can ask a quick question based on my complete lack of understanding ... regarding how dataverse handles file versions ... i understand that it keeps version info about datasets, but when changes are  made at the file level, does it just create a new file? and "versions" of files are literally different files, each with it's own version info?
19:07 pdurbin nightowl313: it's a little confusing. Let me try to explain.
19:07 pdurbin Datasets have versions.
19:08 pdurbin Files are immutable.
19:08 pdurbin So the content of files, the md5, never changes.
19:09 pdurbin But descriptions of files can change. We call this "file metadata".
19:10 pdurbin And when you want to change the description of a file that has been published, you create a new version of the dataset, a draft version.
19:10 jri joined #dataverse
19:10 pdurbin Because files use the same versioning of datasets.
19:11 pdurbin Does that help? I can keep going.
19:12 nightowl313 well, yes, but if I change the actual content of the file and upload that file (with the same name), it just creates a new file? (ie: with a -1 -2, etc to indicate the origin)
19:12 nightowl313 and publish the dataset
19:13 pdurbin Yes but there's also a feature call "file replace" where you can indicate that it's a newer version of the same file.
19:15 nightowl313 i don't see that as an option on the file interface .. how do I accomplish file replace?
19:16 pdurbin The file has to be published already.
19:16 pdurbin Here are the docs on it: http://guides.dataverse.org/en/4.20/user/dataset-management.html#replace-files
19:18 nightowl313 i was looking at that doc ... and i completely missed that ... and apparently I missed the replace option too! I just found it ... so sorry! I've been looking at this for too long
19:19 nightowl313 i think I get it now, thank you so much!
19:21 pdurbin Great!
19:21 nightowl313 and you can still view both versions of the file contents, right?
19:22 pdurbin Oh yes, always. When I say "files are immutable" that's what I mean. You can go back to old versions and download them.
19:23 nightowl313 oh yes, I see I was able to see the file versions, not just the dataset versions ... thanks!
19:25 pdurbin Sure. Because you know, maybe a paper is based on version 2 of a dataset. And you want to go back and reproduce the results.
19:26 nightowl313 yes, for sure .. we are working on preservation and backukp/recovery policies and I knew that I had looked into this before but I just wanted to verify that I could see a file-level version history and view those files
19:27 pdurbin cool
19:27 nightowl313 it is a little confusing because you can see the dataset history and number of files changed, but I just needed to remember to go to the file view of each file to see that history
19:28 nightowl313 but as I was testing, I was just uploading the file again instead of using the "replace" link!
19:28 nightowl313 thanks so much!
19:28 pdurbin No problem. And when do we get you installation on the map? :)
19:29 pdurbin your*
19:29 stefankasberger joined #dataverse
19:29 nightowl313 hopefully soon! we have finished the pilot testing, and are now working on steps and what we need to do to go live ... I do have some results from the testing if you are interested?
19:31 pdurbin I'd be happy to look at them. philip_durbin@harvard.edu
19:32 nightowl313 sure, I will compile and summary and send it ... thank you!
19:37 poikilotherm1 Pdurbin please feel free to request any changes you see necessary for #6974. The more people take a look the better we ensure this is useful for others.
19:45 jri joined #dataverse
19:47 pdurbin poikilotherm1: I would suggest re-wording that line I mentioned earlier.
19:49 poikilotherm1 Do you feel like addding one of your great -+ diffs as a commitable comment? Me being a non-native speaker would appreciate that! :-l
19:50 poikilotherm1 Autocorrection... That should have been a :-)
19:52 pdurbin let me see what I can do... sort of all over the place at the moment
19:53 poikilotherm1 Oh BTW donsizemore I noticed Jenkins has a hickup when doing the report analysis, thus giving a failing report on PRs. Do feel like taking a look at that? No rush...
19:53 poikilotherm1 +you
19:53 poikilotherm1 pdurbin I hope that was not ment literally :-) don't spoil any of that brain on your desk ;-)
19:54 pdurbin I wish I had a desk. I'm typing on the dresser in my bedroom. :(
19:56 poikilotherm1 Oh. Corona Home Office?
19:57 pdurbin yeah
19:58 poikilotherm1 Guys, please stay safe and healthy. I'm really looking forward to seeing you again this week
19:59 pdurbin We'll try. You're coming to the devops session, poikilotherm1?
20:01 nightowl313 left #dataverse
20:35 jri joined #dataverse
20:58 jri joined #dataverse
21:00 pdurbin Heading out. See you tomorrow.
21:00 pdurbin left #dataverse
21:58 nightowl313 joined #dataverse
22:00 nightowl313 for disaster recovery, if we were to restore a dataverse installation that is using S3 for storage of user files, I know I would need to have a backup export of the database, as well as backup of the s3 bucket that I'm using, but i can't just point the new installation to the new s3 .. I have to also change the bucket name in the "dvobject" table for each file as well, right?
22:46 nightowl313 left #dataverse
23:33 nightowl313 joined #dataverse

| Channels | #dataverse index | Today | | Search | Google Search | Plain-Text | plain, newest first | summary

Connect via chat.dataverse.org to discuss Dataverse (dataverse.org, an open source web application for sharing, citing, analyzing, and preserving research data) with users and developers.