IQSS logo

IRC log for #dataverse, 2019-06-28

Connect via chat.dataverse.org to discuss Dataverse (dataverse.org, an open source web application for sharing, citing, analyzing, and preserving research data) with users and developers.

| Channels | #dataverse index | Today | | Search | Google Search | Plain-Text | plain, newest first | summary

All times shown according to UTC.

Time S Nick Message
07:04 icarito[m] joined #dataverse
07:22 poikilotherm joined #dataverse
07:30 jri joined #dataverse
07:47 rigelk joined #dataverse
08:59 jri joined #dataverse
09:02 jri_ joined #dataverse
09:03 poikilotherm pdurbin: See Jenkins and Mr Robot in action...
09:04 poikilotherm https://i.imgur.com/Zyr8950.png
09:04 poikilotherm https://github.com/IQSS/dataverse-kubernetes/pull/74
10:04 juancorr joined #dataverse
10:14 pdurbin poikilotherm: awesome
10:29 poikilotherm I hope for donsizemore to materialize here...
10:34 pdurbin poikilotherm: me too. Let's add your config to dataverse-jenkins. I just created this issue: https://github.com/IQSS/dataverse-jenkins/issues/11
10:35 poikilotherm Yeah, I wanted to play around with it first :-D
10:35 pdurbin sure
10:36 pdurbin And I don't know how we should name these files. config.xml is too broad for what it is. IQSS-dataverse-api-test-su​ite-docker-aio-develop.xml may be a little long.
10:36 poikilotherm maybe the xml should be captured in a subfolder k8s
10:36 poikilotherm I plan to add more jobs
10:37 pdurbin subfolders are fine
10:37 pdurbin We're just getting started.
10:37 poikilotherm Is this repo for documentation / examples or do you plan to do gitops on it?
10:38 pdurbin It's not for examples. It's supposed to represent the actual jobs in use. So that anyone can spin up their own Jenkins server. If you run a fork or whatever, for example.
10:39 pdurbin Please see "How to install Jenkins to build and test Dataverse" at https://github.com/IQSS/dataverse-jenkins/issues/1
10:40 pdurbin I've learned a lot. A month or so ago I had never installed Jenkins. :)
10:45 poikilotherm pdurbin would you be ok with me adding a docker hub user similar to the dataverse-bot?
10:45 poikilotherm To push images automatically, I will need an account over there...
10:55 pdurbin Do you want to use your Docker Hub account for now?
10:56 poikilotherm Meh :-/
10:56 poikilotherm I could, but that means adding my personal credentials to Jenkins
10:56 pdurbin yuck
10:56 poikilotherm I am not very fond of that...
10:57 pdurbin What name would you give it?
10:57 poikilotherm dataverse-bot?
10:57 poikilotherm Same name as github
11:10 pdurbin Sure, sounds fine.
11:11 pdurbin Sorry, distracted. Making breakfast. And my nine year old and I just showed her mom this crazy YouTube video that we keep singing to.
11:12 xarthisius joined #dataverse
11:12 xarthisius joined #dataverse
11:13 pdurbin I'm glad you're getting back to the automated testing stuff. It's so important. The phoenix server caught a regression this week.
12:45 poikilotherm Meh Docker Hub does not allow dashes in usernames :-/
12:52 poikilotherm pdurbin could you grant permission for dataversebot to the repos?
12:52 poikilotherm (On docker hub)
12:59 donsizemore joined #dataverse
12:59 poikilotherm Morning donsizemore :-)
13:01 donsizemore @poikilotherm morning =) i was going to offer to tick off dataverse-kubernetes #11 if i won't step on your toes
13:01 donsizemore also, i suppose it's no longer morning there, but thought that counts!
13:02 donsizemore ^^ the dataverse-java-client job fails because i didn't want to pop any credentials into jenkins IIRC
13:02 poikilotherm By ticking it off you mean you would work on the refactoring in upstream?
13:03 poikilotherm There is an upstream ticket for this, too: https://github.com/IQSS/dataverse/issues/5733
13:04 donsizemore i was thinking i'd just grab the XML and drop it into place? i'll hold off pending refactoring
13:06 poikilotherm I'm buzzled... Grab the XML and drop it? You mean add a workaround?
13:06 donsizemore i think phil just wanted the job config documented
13:07 poikilotherm Are talking about  the same things here?
13:07 poikilotherm +we
13:07 donsizemore p.s. i just noted some of those temperatures in your "heat" wave... it gets hotter than that here =)
13:07 donsizemore we may not be
13:07 poikilotherm I thought you are talking about https://github.com/IQSS/dataverse-kubernetes/issues/11
13:07 donsizemore i mean https://github.com/IQSS/dataverse-jenkins/issues/11
13:08 donsizemore sorry, only on my 4th cup of coffee here
13:08 poikilotherm Ah ah ah! Now that makes sense :-D
13:09 pdurbin_m joined #dataverse
13:09 poikilotherm Feel free :-) That job should be good as is. It will need some additions to the global config and list of plugins
13:09 poikilotherm , too
13:09 poikilotherm Otherwise it would not be redeployable instantly
13:09 pdurbin_m donsizemore: you know what I want. :)
13:10 poikilotherm I'm going to add an enchantment to the card deck "Know what Phil wants"
13:10 pdurbin_m heh
13:11 pdurbin_m I want reproducible builds.
13:11 poikilotherm I'll take that as a cite on the card :-D
13:12 poikilotherm Ok, jokes aside. donsizemore feel free to fire at will
13:12 poikilotherm Be warned though, that I already am working on more jos
13:12 poikilotherm jobs
13:12 pdurbin_m poikilotherm: I'm assuming you have an account on the Jenkins server. I don't. On purpose. :)
13:12 poikilotherm Maybe it might be less work to work on the hole bunch in one go
13:13 poikilotherm pdurbin_m: Oh! Is that so? Why not?
13:13 donsizemore i can update them any time
13:13 poikilotherm As you prefer
13:13 donsizemore hey @pdurbin any way i can help debug https://dataverse.harvard.edu/dataverse/ajps
13:14 donsizemore @pdurbin_m i made @poikilotherm an admin of the jenkins instance hoping i could learn something via osmosis
13:15 poikilotherm donsizemore I was going to ask you about how much steroids the VM has where poor Mr. Robot is going to break stones
13:15 pdurbin_m donsizemore: yes! Please help!!
13:15 poikilotherm But that can wait after debugging the 500
13:15 donsizemore @poikilotherm it currently has 4 cores and 8GB of RAM to match phil's setup at harvard, but i can give it more oomph if need be
13:16 donsizemore @pdurbin_m any chance i can get in your splunk or unprivileged AWS access or...?
13:16 poikilotherm It felt kinda clumsy building the image. But that might be related to stdout capturing
13:16 poikilotherm I don't want to overload ;-)
13:17 donsizemore if it fills up we give it more
13:17 poikilotherm Alright
13:17 poikilotherm What about disk space?
13:17 poikilotherm Images tend to get leave a footprint...
13:18 poikilotherm -get
13:18 pdurbin_m donsizemore: any access you want
13:18 donsizemore i gave it 30G for starters but that was before docker/kubernetes came on the picture
13:19 poikilotherm Yeah, that might be a bit... tight.
13:19 donsizemore whatever can let me read server.log? i can use mandy's api key if we want to try to retrieve info that way
13:19 donsizemore @poikilotherm how much would you suggest?
13:19 poikilotherm Lemme check my docker daemon
13:20 pdurbin_m donsizemore: could you please start by emailing Danny?
13:21 poikilotherm Ok local images ~10GB, but lots of reclaimable. On my minikube ephemeral vm ~10G, but only 50% reclaimable
13:22 poikilotherm So maybe try with 100GB and put monitoring in place?
13:23 poikilotherm Reclaimable is a dangerous thing... Caches speeding up the build are also falling under "reclaimable"...
13:23 pdurbin_m I'm walking kids to camp.
13:24 donsizemore @pdurbin_m done
13:25 donsizemore @poikilotherm i'll give it 100GB once we get AJPS sorted
13:25 poikilotherm Sure, no problem. Its not urgent
13:26 poikilotherm pdurbin: I am outta here in about 30 minutes. Could be so kind to add DockerHub user dataversebot to the IQSS repos for dataverse-k8s and solr-k8s later? Then I have a chance to test things this weekend / Monday morning
13:28 pdurbin donsizemore: thanks! People on Twitter are noticing the instability: https://twitter.com/ronpatz/status/1144494633998639109 :(
13:29 pdurbin poikilotherm: can you please rename the GitHub account to dataversebot (no dash) to match?
13:30 poikilotherm Is that possible?
13:31 pdurbin or make a new account
13:31 poikilotherm Done :-D
13:31 poikilotherm https://github.com/dataversebot
13:36 pdurbin poikilotherm: thanks! And I made a note about your request: https://trello.com/c/e6SHgTnc/3858-could-be-so-kind-to-add-dockerhub-user-dataversebot-to-the-iqss-repos-for-dataverse-k8s-and-solr-k8s-later
13:36 poikilotherm Thx :-)
13:37 xarthisius pdurbin: r2d's Zenodo integration? Yeah I am. What about it?
13:44 pdurbin_m joined #dataverse
13:44 pdurbin_m xarthisius: how do you feel about adding a dataverse.py? :)
13:45 xarthisius I thought I did :) In a slightly different project though ;)
13:47 pdurbin_m xarthisius: right! That's the point! You already wrote the code. I've been thinking it might be somewhat straightforward for you to copy and paste it into dataverse.py.
13:50 pdurbin_m xarthisius: I would be happy to create the issue in repo2docker. :)
13:50 xarthisius some of that code for sure, but there are subtle differences
13:51 xarthisius e.g. we maintain state, so that we can grab location of dataverses once. In r2d that would need happen in `detect`
13:52 xarthisius which means you'd get a request to dataverse installations-json every time someone build a binder...
13:52 xarthisius if you'd port our code directly that is
14:05 pdurbin_m xarthisius: if I find someone who's willing to Port the code are you able to help answer questions?
14:07 xarthisius yes, of course
14:08 pdurbin_m xarthisius: thank you!
14:09 pdurbin_m donsizemore: do you think Andre would be interested?
14:13 pdurbin_m Andrey*
14:20 pdurbin_m xarthisius: oh, can you please link me to the code to port over?
14:32 xarthisius pdurbin_m: https://github.com/whole-tale/girder_wholetale/blob/master/server/lib/dataverse/provider.py
14:33 xarthisius these are Dataverse specific bits, but some of the functionality is general and defined in a class that this specific provider is derived from
14:35 pdurbin_m xarthisius: thanks!
14:41 poikilotherm joined #dataverse
14:41 xarthisius pdurbin_m: is there a way of knowing that a url points to a Dataverse installation (except for checking for a match in installations-json which we do now), or is there a certain range of DOIs that are issued solely for Dataverses?
14:47 donsizemore @pdurbin_m today is andrey's last day
15:06 pdurbin donsizemore: bah. Please tell him I really enjoyed his CoReRe presentation!
15:07 pdurbin xarthisius: I would suggest checking URLs like this: https://demo.dataverse.org/api/info/version
15:09 pdurbin It should return something like this: {"status":"OK","data":{"version"​:"4.14","build":"1326-57a392d"}}
15:20 donsizemore @pdurbin you may now CoReRe all you like https://core2test.irss.unc.edu/
15:21 donsizemore (but you have to roll your neck when you say it)
15:40 pdurbin Heh. "Co Ray Ray!" But I'm getting curl: (7) Failed to connect to core2test.irss.unc.edu port 443: Connection refused
15:45 pdurbin Out of curiousity, what do you have http://guides.dataverse.org/en/4.15/installation/config.html#zipdownloadlimit set to?
15:45 pdurbin andrewSC bricas_ donsizemore juancorr ^^
15:46 donsizemore oh, i need to open the firewall for you all.
15:46 donsizemore "Setting :ZipDownloadLimit not found"
15:46 pdurbin curiosity*
15:47 pdurbin ok, so the default, which is 100 MB if we can trust the docs
15:48 pdurbin donsizemore: do people ever complain that the limit is too low?
16:51 andrewSC joined #dataverse
16:55 bricas_ pdurbin: i see the new note on the dataverse release page on github about pg 9.6 -- does this mean that requirement will not change, or is it still under consideration?
16:58 donsizemore @pdurbin i don't know that we've run into it. let me ask mandy and cheryl
17:01 pdurbin bricas_: well, it's kind of too late for 4.15. PostgreSQL 9.6 is required because of the synax we're using. It would be nice to be more flexible. Not long ago Dataverse ran fine on PostgreSQL 8.4. Maybe 4-6 months ago.
17:01 pdurbin donsizemore: ok, just curious. Thanks.
17:02 bricas_ pdurbin: i understand 4.15 has been minted, but can 4.15.1 use <9.6 syntax and i can upgrade straight to that?
17:03 pdurbin bricas_: well, I don't want to say no. It's more like I don't know. All these SQL upgrade scripts we're writing are now being tracked in the "flyway_schema_history" table of your database: http://phoenix.dataverse.org/schemaspy/latest/tables/flyway_schema_history.html
17:04 pdurbin ... and I don't really know what it means to try to edit a script that's already in that table from the past.
17:05 pdurbin We're still getting used to Flyway and it's a little painful to be honest. We haven't studied it very closely. The goal was to make upgrades for sysadmins easier. No more having to run SQL scripts on the side. They get run automatically on deployment.
17:07 bricas_ understood. just wondering what to tell others, like, we're stuck on 5.14 until we upgrade pg, which means a non-base apt repo... or, maybe it'll work in a future release if we just wait?
17:15 pdurbin Honestly, I would suggest upgrading Postgres, even if it means a non-base apt repo. That's what we all do on CentOS. You could try Posgres 10, maybe with donsizemore's blessing. :)
17:15 pdurbin donsizemore ran Shib 3 before anyone else. He's our pioneer.
17:15 donsizemore @bricas_ i'd want to run whatever harvard's running...
17:17 pdurbin Harvard Dataverse uses Amazon's "postgres as a service" so our upgrades are tied to that. I think we had to get off postgres 9.3 recently because it was going away.
17:18 donsizemore @pdurbin it didn't go away, it just stopped receiving security updates. we're migrating to our new Dataverse digs (and Postgres-9.6) Real Soon Now
17:18 pdurbin "Amazon RDS will be discontinuing support for version 9.3 of the PostgreSQL database in RDS, starting on September 5th, 2018." https://github.com/IQSS/dataverse/issues/4709
17:18 pdurbin donsizemore: are you on RDS too?
17:19 donsizemore no, we're running 9.3 from the postgres repo
17:19 pdurbin ok
17:19 donsizemore (it was the newest stable version when we migrated)
17:19 pdurbin bricas_: I'm not sure if this is helping you. :)
17:19 bricas_ heh
17:19 donsizemore @bricas_ how big is your DB?
17:19 bricas_ very small
17:19 pdurbin I mean, eventually we'll all be off Postgres 9, right? 5 or 10 years from now, if not sooner. :)
17:20 donsizemore ours is 144MB in postgres' custom dump format. takes about 3 minutes to import to a test box
17:21 bricas_ 222K Jun 28 12:00 dvndb.gz
17:21 bricas_ :)
17:22 bricas_ i will say that it is unlikely pg 9.5 will ever be supported so we should plan on 9.6
17:22 donsizemore should be pretty easy to dump and reload as a test?
17:22 bricas_ typically custom repos aren't discouraged, but we have been bit before when an os upgrade didn't quite go as planned
17:23 donsizemore i'd never want to do an in-place upgrade on a production box
17:23 donsizemore i've pushed our production database around to test machines like a bad fruitcake, and thankfully never encountered import errors
17:24 bricas_ i've been doing super fun in place upgrades of our ubuntu 14.04 boxes since january (14.04->16.04->18.04). got them all done by end of april luckily.
17:25 donsizemore maybe i got burned enough during the 8.04/10.04/12.04 days that i haven't tried an in-place since then
17:25 pdurbin I always just spin up a new Digital Ocean droplet for a new Ubuntu LTS for my home stuff. I guess I'm just paranoid from many years of using CentOS.
17:27 bricas_ it kind of a shit-show. the upgrade from 14.04->16.04 will die because they somehow messed up the mysql 5.6 -> 5.7 upgrade which means you have to restore all dbs from disk
17:28 bricas_ anyway, another 4 years until i worry about it again ;)
17:28 pdurbin :)
17:31 donsizemore @pdurbin dumb question?
17:32 pdurbin hit me
17:32 donsizemore mandy's having trouble publishing up yonder, so i tried via the API per http://guides.dataverse.org/en/latest/api/sword.html#publish-a-dataset
17:33 donsizemore i cobble together the command, but dataverse.harvard challenges: "Enter host password for user '<api-token>':
17:34 donsizemore if this is expected, i'll have her come in and enter her password, but...
17:34 pdurbin no, a password shouldn't be required
17:34 pdurbin an api token should work
17:34 pdurbin you could also try the native api: http://guides.dataverse.org/en/4.15/api/native-api.html#publish-a-dataset
17:35 donsizemore oh. pffft. sword came back from a search
17:35 pdurbin but that sword version should work too
17:35 pdurbin clicking publish doesn't work?
17:36 donsizemore my suspicion is it's datacite but i'll try via native. one sec
17:38 pdurbin anything different about the dataset?
17:39 donsizemore native API threw a 404 using the DOI as the ID. i'll ask her to sneaky-sneaky me the database ID
17:40 donsizemore she said what had happened was "it says that it is updating or registering file identifiers and then a little spinny wheel might pop up, but then it just kicks me back to the already published v1.0 If i go to edit metadata, my draft is still there.... it just won't update to v1.1 with the updated metadata"
17:42 pdurbin donsizemore: can you please ask her to open a ticket by emailing support@dataverse.harvard.edu ?
17:43 donsizemore yis. tried native API with database ID and got a 503, will keep trying
17:44 donsizemore mine was more of a usage question or to see if you'd say "yeah, dat broke"
17:45 donsizemore native API throws a 404 with DB ID. support it is
17:50 pdurbin The APIs should work fine. Clicking "Publish" should work fine too. I don't know what's going on.
17:52 donsizemore i blame AWS. i got the API to respond "WORK_IN_PROGRESS" but mandy says it's still at 1.0
17:57 pdurbin I got nuthin. Sorry.
18:04 donsizemore i think it's trying to tell us to leave early for andrey's happy hour
18:05 pdurbin probably, and thanks for opening https://help.hmdc.harvard.edu/Ticket/Display.html?id=278072
18:06 donsizemore Dwayne!
18:09 donsizemore ok, i'm out. have a great weekend!
19:35 andrewSC joined #dataverse
19:39 pdurbin I'm out too. Have a good weekend, everyone.
19:39 pdurbin left #dataverse

| Channels | #dataverse index | Today | | Search | Google Search | Plain-Text | plain, newest first | summary

Connect via chat.dataverse.org to discuss Dataverse (dataverse.org, an open source web application for sharing, citing, analyzing, and preserving research data) with users and developers.