I noticed CRAN is now doing checks against Apple M1, and some packages are
failing including a dependency I use. Is building on M1 now a requirement, or can the check be ignored? If it's a requirement, how can one test it out? Travers [[alternative HTML version deleted]] ______________________________________________ [hidden email] mailing list https://stat.ethz.ch/mailman/listinfo/r-devel |
On 22/02/2021 08:30, Travers Ching wrote:
> I noticed CRAN is now doing checks against Apple M1, and some packages are > failing including a dependency I use. I don't know what this refers to: M1 Mac CRAN checks are planned but AFAICS not yet included in the main results tables. OTOH, 'Additional issues' on M1 Mac have been reported on the results pages since early December. > Is building on M1 now a requirement, or can the check be ignored? If it's a > requirement, how can one test it out? 'requirement' for what? I am not aware of any CRAN package for which 'R CMD build' does not work on an M1 Mac. *Checking* might need an M1 Mac machine. CRAN has only been notifying issues which can easily be corrected without access to M1 hardware (such as using suggested packages unconditionally or using optional capabilities without checking). > Travers > > [[alternative HTML version deleted]] Please do re-read the posting guide (and 'Writing R Extensions'). Also, this is not r-package-devel .... -- Brian D. Ripley, [hidden email] Emeritus Professor of Applied Statistics, University of Oxford ______________________________________________ [hidden email] mailing list https://stat.ethz.ch/mailman/listinfo/r-devel |
Hi Prof Ripley,
Here is the automated message from CRAN which I thought meant needing to fix an M1 issue: "The auto-check found additional issues for the *last* version released on CRAN: M1mac <https://www.stats.ox.ac.uk/pub/bdr/M1mac/stringfish.out> CRAN incoming checks do not test for these additional issues and you will need an appropriately instrumented build of R to reproduce these. Hence please reply-all and explain: Have these been fixed? " However, RcppParallel (a dependency) isn't building on M1: https://www.stats.ox.ac.uk/pub/bdr/M1mac/RcppParallel.out If I understand you correctly, I can ignore the M1 "Additional issues" until official R support? Thank you, Travers On Mon, Feb 22, 2021 at 11:25 PM Prof Brian Ripley <[hidden email]> wrote: > On 22/02/2021 08:30, Travers Ching wrote: > > I noticed CRAN is now doing checks against Apple M1, and some packages > are > > failing including a dependency I use. > > I don't know what this refers to: M1 Mac CRAN checks are planned but > AFAICS not yet included in the main results tables. > > OTOH, 'Additional issues' on M1 Mac have been reported on the results > pages since early December. > > > Is building on M1 now a requirement, or can the check be ignored? If > it's a > > requirement, how can one test it out? > > 'requirement' for what? > > I am not aware of any CRAN package for which 'R CMD build' does not work > on an M1 Mac. > > *Checking* might need an M1 Mac machine. CRAN has only been notifying > issues which can easily be corrected without access to M1 hardware (such > as using suggested packages unconditionally or using optional > capabilities without checking). > > > Travers > > > > [[alternative HTML version deleted]] > > Please do re-read the posting guide (and 'Writing R Extensions'). > Also, this is not r-package-devel .... > > -- > Brian D. Ripley, [hidden email] > Emeritus Professor of Applied Statistics, University of Oxford > [[alternative HTML version deleted]] ______________________________________________ [hidden email] mailing list https://stat.ethz.ch/mailman/listinfo/r-devel |
In reply to this post by Travers Ching
If there was a response to the "how can I test it out" part of this
question then I missed it. Can anyone point to a Win-builder like site for testing on M1mac, or to the M1mac results from testing packages already on CRAN? They still do not seem to be on the CRAN daily site. Even a link to the 'Additional issues' on M1 Mac on the results pages would be helpful because it does not seem to be in an obvious place. I am trying to respond to a demand to relax or remove some package testing that fails because M1mac gives results outside my specified tolerances. The tests in question (in package EvalEst) have been used since very early R versions (0.16 circa 1995), and used on Splus prior to that. There has been a need to adjust tolerances occasionally, but they have been stable for a long time (more than 20 years I believe). Since these tests date from a time when simple double precision was the norm, the tolerances are already fairly relaxed so I hesitate to adjust them with actually examining the results. Paul Gilbert On 2021-02-22 3:30 a.m., Travers Ching wrote: > I noticed CRAN is now doing checks against Apple M1, and some packages are > failing including a dependency I use. > > Is building on M1 now a requirement, or can the check be ignored? If it's a > requirement, how can one test it out? > > Travers > > [[alternative HTML version deleted]] > > ______________________________________________ > [hidden email] mailing list > https://stat.ethz.ch/mailman/listinfo/r-devel > ______________________________________________ [hidden email] mailing list https://stat.ethz.ch/mailman/listinfo/r-devel |
Dear Gilbert,
it would obviously great to have something like Win-builder/R-hub for the architecture of the M1 macs. I had an issue with the RcppCWB package I maintain in January and I could not get hold of a M1 Mac mini quickly. So I resorted to hiring one with MacStadium (https://www.macstadium.com/). That worked nicely and I had a testing environment within a few hours. But it is an expensive solution (~ 100$ per month)! In the meantime I got myself a Mac mini... Great machine, but obviously not everybody should not be forced to buy one. So it would be great to learn whether there are plans to offer testing environments for Apple M1. Andreas Blätte Am 28.02.21, 16:50 schrieb "R-devel im Auftrag von Paul Gilbert" <[hidden email] im Auftrag von [hidden email]>: If there was a response to the "how can I test it out" part of this question then I missed it. Can anyone point to a Win-builder like site for testing on M1mac, or to the M1mac results from testing packages already on CRAN? They still do not seem to be on the CRAN daily site. Even a link to the 'Additional issues' on M1 Mac on the results pages would be helpful because it does not seem to be in an obvious place. I am trying to respond to a demand to relax or remove some package testing that fails because M1mac gives results outside my specified tolerances. The tests in question (in package EvalEst) have been used since very early R versions (0.16 circa 1995), and used on Splus prior to that. There has been a need to adjust tolerances occasionally, but they have been stable for a long time (more than 20 years I believe). Since these tests date from a time when simple double precision was the norm, the tolerances are already fairly relaxed so I hesitate to adjust them with actually examining the results. Paul Gilbert On 2021-02-22 3:30 a.m., Travers Ching wrote: > I noticed CRAN is now doing checks against Apple M1, and some packages are > failing including a dependency I use. > > Is building on M1 now a requirement, or can the check be ignored? If it's a > requirement, how can one test it out? > > Travers > > [[alternative HTML version deleted]] > > ______________________________________________ > [hidden email] mailing list > https://stat.ethz.ch/mailman/listinfo/r-devel > ______________________________________________ [hidden email] mailing list https://stat.ethz.ch/mailman/listinfo/r-devel ______________________________________________ [hidden email] mailing list https://stat.ethz.ch/mailman/listinfo/r-devel |
In reply to this post by Paul Gilbert-2
Paul,
this is being worked on. As you can imagine testing over 17,000 package in a M1 Mac mini isn't quite trivial. The first priority was to get the nightly R builds to work. Second was to get CRAN package builds to work. Third is to provide checks. The first two have finished last week and the checks have been running for the past two days. Unfortunately, some pieces (like XQuartz) are still not quite stable so it takes more manual interventions than one would expect. We are at close to 16k packags checked, so we're getting there. As for EvalEst the check has finished so I have: Running ‘dse2tstgd2.R’ [13s/14s] Running the tests in ‘tests/dse2tstgd2.R’ failed. Last 13 lines of output: > ok <- fuzz.large > error > if (!ok) {if (is.na(max.error)) max.error <- error + else max.error <- max(error, max.error)} > all.ok <- all.ok & ok > {if (ok) cat("ok\n") else cat("failed! error= ", error,"\n") } ok > > cat("All Brief User Guide example tests part 2 completed") All Brief User Guide example tests part 2 completed> if (all.ok) cat(" OK\n") else + cat(", some FAILED! max.error = ", max.error,"\n") , some FAILED! max.error = 1.065814e-14 > > if (!all.ok) stop("Some tests FAILED") Error: Some tests FAILED Execution halted when I run it by hand I get ok for all but: Guide part 2 test 10... failed! error= 1.065814e-14 > sum(fc1$forecastCov[[1]]) [1] 14.933660144821400806 > sum(fc2$forecastCov[[1]]) [1] 14.933660144821400806 > sum(fc2$forecastCov.zero) [1] 31.654672476928304548 > sum(fc2$forecastCov.trend) [1] 18.324461923341957004 > c(14.933660144821400806 - sum(fc1$forecastCov[[1]]), + 14.933660144821400806 - sum(fc2$forecastCov[[1]]), + 31.654672476928297442 - sum(fc2$forecastCov.zero), + 18.324461923341953451 - sum(fc2$forecastCov.trend) ) [1] 0.0000000000000000000e+00 0.0000000000000000000e+00 -1.0658141036401502788e-14 [4] -3.5527136788005009294e-15 I hope this helps you to track it down. Cheers, Simon > On Mar 1, 2021, at 4:50 AM, Paul Gilbert <[hidden email]> wrote: > > If there was a response to the "how can I test it out" part of this question then I missed it. Can anyone point to a Win-builder like site for testing on M1mac, or to the M1mac results from testing packages already on CRAN? They still do not seem to be on the CRAN daily site. Even a link to the 'Additional issues' on M1 Mac on the results pages would be helpful because it does not seem to be in an obvious place. I am trying to respond to a demand to relax or remove some package testing that fails because M1mac gives results outside my specified tolerances. > > The tests in question (in package EvalEst) have been used since very early R versions (0.16 circa 1995), and used on Splus prior to that. There has been a need to adjust tolerances occasionally, but they have been stable for a long time (more than 20 years I believe). Since these tests date from a time when simple double precision was the norm, the tolerances are already fairly relaxed so I hesitate to adjust them with actually examining the results. > > Paul Gilbert > > On 2021-02-22 3:30 a.m., Travers Ching wrote: >> I noticed CRAN is now doing checks against Apple M1, and some packages are >> failing including a dependency I use. >> Is building on M1 now a requirement, or can the check be ignored? If it's a >> requirement, how can one test it out? >> Travers >> [[alternative HTML version deleted]] >> ______________________________________________ >> [hidden email] mailing list >> https://stat.ethz.ch/mailman/listinfo/r-devel >> > > ______________________________________________ > [hidden email] mailing list > https://stat.ethz.ch/mailman/listinfo/r-devel > ______________________________________________ [hidden email] mailing list https://stat.ethz.ch/mailman/listinfo/r-devel |
Simon,
Yes, I can imagine it is not trivial testing. I hope you have a stack of minis in a cluster. It looks like a trivial transgression, max.error = 1.065814e-14 where my tolerance is set at 1e-14. Many of the other tests already have a larger tolerance. Very possibly it is not actually the tolerance that is the problem. The test value itself is just determined by a run on another machine, so that may not be in the middle of the result distribution. I'll fix it shortly. Thanks, Paul On 2021-02-28 4:23 p.m., Simon Urbanek wrote: > Paul, > > this is being worked on. As you can imagine testing over 17,000 package in a M1 Mac mini isn't quite trivial. The first priority was to get the nightly R builds to work. Second was to get CRAN package builds to work. Third is to provide checks. The first two have finished last week and the checks have been running for the past two days. Unfortunately, some pieces (like XQuartz) are still not quite stable so it takes more manual interventions than one would expect. We are at close to 16k packags checked, so we're getting there. > > As for EvalEst the check has finished so I have: > > Running ‘dse2tstgd2.R’ [13s/14s] > Running the tests in ‘tests/dse2tstgd2.R’ failed. > Last 13 lines of output: > > ok <- fuzz.large > error > > if (!ok) {if (is.na(max.error)) max.error <- error > + else max.error <- max(error, max.error)} > > all.ok <- all.ok & ok > > {if (ok) cat("ok\n") else cat("failed! error= ", error,"\n") } > ok > > > > cat("All Brief User Guide example tests part 2 completed") > All Brief User Guide example tests part 2 completed> if (all.ok) cat(" OK\n") else > + cat(", some FAILED! max.error = ", max.error,"\n") > , some FAILED! max.error = 1.065814e-14 > > > > if (!all.ok) stop("Some tests FAILED") > Error: Some tests FAILED > Execution halted > > when I run it by hand I get ok for all but: > > Guide part 2 test 10... > failed! error= 1.065814e-14 > >> sum(fc1$forecastCov[[1]]) > [1] 14.933660144821400806 >> sum(fc2$forecastCov[[1]]) > [1] 14.933660144821400806 >> sum(fc2$forecastCov.zero) > [1] 31.654672476928304548 >> sum(fc2$forecastCov.trend) > [1] 18.324461923341957004 >> c(14.933660144821400806 - sum(fc1$forecastCov[[1]]), > + 14.933660144821400806 - sum(fc2$forecastCov[[1]]), > + 31.654672476928297442 - sum(fc2$forecastCov.zero), > + 18.324461923341953451 - sum(fc2$forecastCov.trend) ) > [1] 0.0000000000000000000e+00 0.0000000000000000000e+00 -1.0658141036401502788e-14 > [4] -3.5527136788005009294e-15 > > I hope this helps you to track it down. > > Cheers, > Simon > > >> On Mar 1, 2021, at 4:50 AM, Paul Gilbert <[hidden email]> wrote: >> >> If there was a response to the "how can I test it out" part of this question then I missed it. Can anyone point to a Win-builder like site for testing on M1mac, or to the M1mac results from testing packages already on CRAN? They still do not seem to be on the CRAN daily site. Even a link to the 'Additional issues' on M1 Mac on the results pages would be helpful because it does not seem to be in an obvious place. I am trying to respond to a demand to relax or remove some package testing that fails because M1mac gives results outside my specified tolerances. >> >> The tests in question (in package EvalEst) have been used since very early R versions (0.16 circa 1995), and used on Splus prior to that. There has been a need to adjust tolerances occasionally, but they have been stable for a long time (more than 20 years I believe). Since these tests date from a time when simple double precision was the norm, the tolerances are already fairly relaxed so I hesitate to adjust them with actually examining the results. >> >> Paul Gilbert >> >> On 2021-02-22 3:30 a.m., Travers Ching wrote: >>> I noticed CRAN is now doing checks against Apple M1, and some packages are >>> failing including a dependency I use. >>> Is building on M1 now a requirement, or can the check be ignored? If it's a >>> requirement, how can one test it out? >>> Travers >>> [[alternative HTML version deleted]] >>> ______________________________________________ >>> [hidden email] mailing list >>> https://stat.ethz.ch/mailman/listinfo/r-devel >>> >> >> ______________________________________________ >> [hidden email] mailing list >> https://stat.ethz.ch/mailman/listinfo/r-devel >> > ______________________________________________ [hidden email] mailing list https://stat.ethz.ch/mailman/listinfo/r-devel |
Free forum by Nabble | Edit this page |