Your comments

I tried again today, and left it waiting for about 2 minutes and it succeeded!!

I guess I thought it was hanging, but it just takes an unusually long time. This is great! I had already figured I'd just comment the imageGet section out and change the template to use cover.jpg (which mylar provides) instead.. Maybe even add an automator script to crop it to a square... But this is great! Finally!


Yay!


Now I just need to build a Hazel script to automatically launch the bash script whenever a cvinfo file is created in my comics folder.

here's usage for curl in Mac


$ curl --help

Usage: curl [options...] <url>

Options: (H) means HTTP/HTTPS only, (F) means FTP only

--anyauth Pick "any" authentication method (H)

-a, --append Append to target file when uploading (F/SFTP)

--basic Use HTTP Basic Authentication (H)

--cacert FILE CA certificate to verify peer against (SSL)

--capath DIR CA directory to verify peer against (SSL)

-E, --cert CERT[:PASSWD] Client certificate file and password (SSL)

--cert-status Verify the status of the server certificate (SSL)

--cert-type TYPE Certificate file type (DER/PEM/ENG) (SSL)

--ciphers LIST SSL ciphers to use (SSL)

--compressed Request compressed response (using deflate or gzip)

-K, --config FILE Read config from FILE

--connect-timeout SECONDS Maximum time allowed for connection

-C, --continue-at OFFSET Resumed transfer OFFSET

-b, --cookie STRING/FILE Read cookies from STRING/FILE (H)

-c, --cookie-jar FILE Write cookies to FILE after operation (H)

--create-dirs Create necessary local directory hierarchy

--crlf Convert LF to CRLF in upload

--crlfile FILE Get a CRL list in PEM format from the given file

-d, --data DATA HTTP POST data (H)

--data-raw DATA HTTP POST data, '@' allowed (H)

--data-ascii DATA HTTP POST ASCII data (H)

--data-binary DATA HTTP POST binary data (H)

--data-urlencode DATA HTTP POST data url encoded (H)

--delegation STRING GSS-API delegation permission

--digest Use HTTP Digest Authentication (H)

--disable-eprt Inhibit using EPRT or LPRT (F)

--disable-epsv Inhibit using EPSV (F)

--dns-servers DNS server addrs to use: 1.1.1.1;2.2.2.2

--dns-interface Interface to use for DNS requests

--dns-ipv4-addr IPv4 address to use for DNS requests, dot notation

--dns-ipv6-addr IPv6 address to use for DNS requests, dot notation

-D, --dump-header FILE Write the headers to FILE

--egd-file FILE EGD socket path for random data (SSL)

--engine ENGINE Crypto engine (use "--engine list" for list) (SSL)

-f, --fail Fail silently (no output at all) on HTTP errors (H)

--false-start Enable TLS False Start.

-F, --form CONTENT Specify HTTP multipart POST data (H)

--form-string STRING Specify HTTP multipart POST data (H)

--ftp-account DATA Account data string (F)

--ftp-alternative-to-user COMMAND String to replace "USER [name]" (F)

--ftp-create-dirs Create the remote dirs if not present (F)

--ftp-method [MULTICWD/NOCWD/SINGLECWD] Control CWD usage (F)

--ftp-pasv Use PASV/EPSV instead of PORT (F)

-P, --ftp-port ADR Use PORT with given address instead of PASV (F)

--ftp-skip-pasv-ip Skip the IP address for PASV (F)

--ftp-pret Send PRET before PASV (for drftpd) (F)

--ftp-ssl-ccc Send CCC after authenticating (F)

--ftp-ssl-ccc-mode ACTIVE/PASSIVE Set CCC mode (F)

--ftp-ssl-control Require SSL/TLS for FTP login, clear for transfer (F)

-G, --get Send the -d data with a HTTP GET (H)

-g, --globoff Disable URL sequences and ranges using {} and []

-H, --header LINE Pass custom header LINE to server (H)

-I, --head Show document info only

-h, --help This help text

--hostpubmd5 MD5 Hex-encoded MD5 string of the host public key. (SSH)

-0, --http1.0 Use HTTP 1.0 (H)

--http1.1 Use HTTP 1.1 (H)

--http2 Use HTTP 2 (H)

--ignore-content-length Ignore the HTTP Content-Length header

-i, --include Include protocol headers in the output (H/F)

-k, --insecure Allow connections to SSL sites without certs (H)

--interface INTERFACE Use network INTERFACE (or address)

-4, --ipv4 Resolve name to IPv4 address

-6, --ipv6 Resolve name to IPv6 address

-j, --junk-session-cookies Ignore session cookies read from file (H)

--keepalive-time SECONDS Wait SECONDS between keepalive probes

--key KEY Private key file name (SSL/SSH)

--key-type TYPE Private key file type (DER/PEM/ENG) (SSL)

--krb LEVEL Enable Kerberos with security LEVEL (F)

--libcurl FILE Dump libcurl equivalent code of this command line

--limit-rate RATE Limit transfer speed to RATE

-l, --list-only List only mode (F/POP3)

--local-port RANGE Force use of RANGE for local port numbers

-L, --location Follow redirects (H)

--location-trusted Like '--location', and send auth to other hosts (H)

--login-options OPTIONS Server login options (IMAP, POP3, SMTP)

-M, --manual Display the full manual

--mail-from FROM Mail from this address (SMTP)

--mail-rcpt TO Mail to this/these addresses (SMTP)

--mail-auth AUTH Originator address of the original email (SMTP)

--max-filesize BYTES Maximum file size to download (H/F)

--max-redirs NUM Maximum number of redirects allowed (H)

-m, --max-time SECONDS Maximum time allowed for the transfer

--metalink Process given URLs as metalink XML file

--negotiate Use HTTP Negotiate (SPNEGO) authentication (H)

-n, --netrc Must read .netrc for user name and password

--netrc-optional Use either .netrc or URL; overrides -n

--netrc-file FILE Specify FILE for netrc

-:, --next Allows the following URL to use a separate set of options

--no-alpn Disable the ALPN TLS extension (H)

-N, --no-buffer Disable buffering of the output stream

--no-keepalive Disable keepalive use on the connection

--no-npn Disable the NPN TLS extension (H)

--no-sessionid Disable SSL session-ID reusing (SSL)

--noproxy List of hosts which do not use proxy

--ntlm Use HTTP NTLM authentication (H)

--oauth2-bearer TOKEN OAuth 2 Bearer Token (IMAP, POP3, SMTP)

-o, --output FILE Write to FILE instead of stdout

--pass PASS Pass phrase for the private key (SSL/SSH)

--path-as-is Do not squash .. sequences in URL path

--pinnedpubkey FILE Public key (PEM/DER) to verify peer against (OpenSSL/GnuTLS/NSS/wolfSSL/CyaSSL/GSKit only)

--post301 Do not switch to GET after following a 301 redirect (H)

--post302 Do not switch to GET after following a 302 redirect (H)

--post303 Do not switch to GET after following a 303 redirect (H)

-#, --progress-bar Display transfer progress as a progress bar

--proto PROTOCOLS Enable/disable PROTOCOLS

--proto-redir PROTOCOLS Enable/disable PROTOCOLS on redirect

-x, --proxy [PROTOCOL://]HOST[:PORT] Use proxy on given port

--proxy-anyauth Pick "any" proxy authentication method (H)

--proxy-basic Use Basic authentication on the proxy (H)

--proxy-digest Use Digest authentication on the proxy (H)

--proxy-negotiate Use HTTP Negotiate (SPNEGO) authentication on the proxy (H)

--proxy-ntlm Use NTLM authentication on the proxy (H)

--proxy-service-name NAME SPNEGO proxy service name

--service-name NAME SPNEGO service name

-U, --proxy-user USER[:PASSWORD] Proxy user and password

--proxy1.0 HOST[:PORT] Use HTTP/1.0 proxy on given port

-p, --proxytunnel Operate through a HTTP proxy tunnel (using CONNECT)

--pubkey KEY Public key file name (SSH)

-Q, --quote CMD Send command(s) to server before transfer (F/SFTP)

--random-file FILE File for reading random data from (SSL)

-r, --range RANGE Retrieve only the bytes within RANGE

--raw Do HTTP "raw"; no transfer decoding (H)

-e, --referer Referer URL (H)

-J, --remote-header-name Use the header-provided filename (H)

-O, --remote-name Write output to a file named as the remote file

--remote-name-all Use the remote file name for all URLs

-R, --remote-time Set the remote file's time on the local output

-X, --request COMMAND Specify request command to use

--resolve HOST:PORT:ADDRESS Force resolve of HOST:PORT to ADDRESS

--retry NUM Retry request NUM times if transient problems occur

--retry-delay SECONDS Wait SECONDS between retries

--retry-max-time SECONDS Retry only within this period

--sasl-ir Enable initial response in SASL authentication

-S, --show-error Show error. With -s, make curl show errors when they occur

-s, --silent Silent mode (don't output anything)

--socks4 HOST[:PORT] SOCKS4 proxy on given host + port

--socks4a HOST[:PORT] SOCKS4a proxy on given host + port

--socks5 HOST[:PORT] SOCKS5 proxy on given host + port

--socks5-hostname HOST[:PORT] SOCKS5 proxy, pass host name to proxy

--socks5-gssapi-service NAME SOCKS5 proxy service name for GSS-API

--socks5-gssapi-nec Compatibility with NEC SOCKS5 server

-Y, --speed-limit RATE Stop transfers below RATE for 'speed-time' secs

-y, --speed-time SECONDS Trigger 'speed-limit' abort after SECONDS (default: 30)

--ssl Try SSL/TLS (FTP, IMAP, POP3, SMTP)

--ssl-reqd Require SSL/TLS (FTP, IMAP, POP3, SMTP)

-2, --sslv2 Use SSLv2 (SSL)

-3, --sslv3 Use SSLv3 (SSL)

--ssl-allow-beast Allow security flaw to improve interop (SSL)

--stderr FILE Where to redirect stderr (use "-" for stdout)

--tcp-nodelay Use the TCP_NODELAY option

-t, --telnet-option OPT=VAL Set telnet option

--tftp-blksize VALUE Set TFTP BLKSIZE option (must be >512)

-z, --time-cond TIME Transfer based on a time condition

-1, --tlsv1 Use => TLSv1 (SSL)

--tlsv1.0 Use TLSv1.0 (SSL)

--tlsv1.1 Use TLSv1.1 (SSL)

--tlsv1.2 Use TLSv1.2 (SSL)

--trace FILE Write a debug trace to FILE

--trace-ascii FILE Like --trace, but without hex output

--trace-time Add time stamps to trace/verbose output

--tr-encoding Request compressed transfer encoding (H)

-T, --upload-file FILE Transfer FILE to destination

--url URL URL to work with

-B, --use-ascii Use ASCII/text transfer

-u, --user USER[:PASSWORD] Server user and password

--tlsuser USER TLS username

--tlspassword STRING TLS password

--tlsauthtype STRING TLS authentication type (default: SRP)

--unix-socket FILE Connect through this Unix domain socket

-A, --user-agent STRING Send User-Agent STRING to server (H)

-v, --verbose Make the operation more talkative

-V, --version Show version number and quit

-w, --write-out FORMAT Use output FORMAT after completion

--xattr Store metadata in extended file attributes

-q Disable .curlrc (must be first parameter)


hmm, seems to just hang there... I CTRL-C after 30s.. Should I wait longer?


$ /Users/Telecart/Downloads/Ubooquity/PageBuilder/imageGet.sh --comic "Action Comics" -r

Comic: Action Comics

Result:

Getting series ID

^C


...

This command does seem to bring back the SERP, so I don't think it's the wget itself...


wget -e robots=off --user-agent "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_11_5) AppleWebKit/601.6.17 (KHTML, like Gecko) Version/9.1.1 Safari/601.6.17" -qO - "https://www.google.com/search?client=safari&rls=en&q='action+comics'+inurl:series+intitle:Digital&ie=UTF-8&oe=UTF=8"

'action comics' inurl:series intitle:Digital - Google Search (function(){window.google={kEI:'EJD5V6WcOIHne5qzubgE',kEXPI:'20782,750721,1351633,1351949,1352019,3700250,3700400,4029370,4029815,4031109'action comics' inurl:series intitle:Digital - Google Search (function(){window.google={kEI:'EJD5V6WcOIHne5qzubgE',kEXPI:'20782,750721,1351633,1351949,1352019,3700250,3700400,4029370,4029815,4031109....'action comics' inurl:series intitle:Digital - Google Search (function(){window.google={kEI:'EJD5V6WcOIHne5qzubgE',kEXPI:'20782,750721,1351633,1351949,1352019,3700250,3700400,4029370,4029815,4031109....'action comics' inurl:series intitle:Digital - Google Search (function(){window.google={kEI:'EJD5V6WcOIHne5qzubgE',kEXPI:'20782,750721,1351633,1351949,1352019,3700250,3700400,4029370,4029815,4031109....'action comics' inurl:series intitle:Digital - Google Search (function(){window.google={kEI:'EJD5V6WcOIHne5qzubgE',kEXPI:'20782,750721,1351633,1351949,1352019,3700250,3700400,4029370,4029815,4031109....'action comics' inurl:series intitle:Digital - Google Search (function(){window.google={kEI:'EJD5V6WcOIHne5qzubgE',kEXPI:'20782,750721,1351633,1351949,1352019,3700250,3700400,4029370,4029815,4031109....'action comics' inurl:series intitle:Digital - Google Search (function(){window.google={kEI:'EJD5V6WcOIHne5qzubgE',kEXPI:'20782,750721,1351633,1351949,1352019,3700250,3700400,4029370,4029815,4031109....'action comics' inurl:series intitle:Digital - Google Search (function(){window.google={kEI:'EJD5V6WcOIHne5qzubgE',kEXPI:'20782,750721,1351633,1351949,1352019,3700250,3700400,4029370,4029815,4031109....'action comics' inurl:series intitle:Digital - Google Search (function(){window.google={kEI:'EJD5V6WcOIHne5qzubgE',kEXPI:'20782,750721,1351633,1351949,1352019,3700250,3700400,4029370,4029815,4031109....'action comics' inurl:series intitle:Digital - Google Search (function(){window.google={kEI:'EJD5V6WcOIHne5qzubgE',kEXPI:'20782,750721,1351633,1351949,1352019,3700250,3700400,4029370,4029815,4031109....'action comics' inurl:series intitle:Digital - Google Search (function(){window.google={kEI:'EJD5V6WcOIHne5qzubgE',kEXPI:'20782,750721,1351633,1351949,1352019,3700250,3700400,4029370,4029815,4031109....'action comics' inurl:series intitle:Digital - Google Search (function(){window.google={kEI:'EJD5V6WcOIHne5qzubgE',kEXPI:'20782,750721,1351633,1351949,1352019,3700250,3700400,4029370,4029815,4031109....'action comics' inurl:series intitle:Digital - Google Search (function(){window.google={kEI:'EJD5V6WcOIHne5qzubgE',kEXPI:'20782,750721,1351633,1351949,1352019,3700250,3700400,4029370,4029815,4031109....'action comics' inurl:series intitle:Digital - Google Search (function(){window.google={kEI:'EJD5V6WcOIHne5qzubgE',kEXPI:'20782,750721,1351633,1351949,1352019,3700250,3700400,4029370,4029815,4031109....'action comics' inurl:series intitle:Digital - Google Search (function(){window.google={kEI:'EJD5V6WcOIHne5qzubgE',kEXPI:'20782,750721,1351633,1351949,1352019,3700250,3700400,4029370,4029815,4031109<!doctype html><html itemscope="" itemtype="http://schema.org/SearchResultsPage" lang="en"><head><meta content="/images/branding/googleg/1x/googleg_standard_color_128dp.png" itemprop="image"><link href="/images/branding/product/ico/googleg_lodp.ico" rel="shortcut icon"><meta content="origin" id="mref" name="referrer"><title>'action comics' inurl:series intitle:Digital - Google Search</title> <script>(function(){window.google={kEI:'EJD5V6WcOIHne5qzubgE',kEXPI:'20782,750721,1351633,1351949,1352019,3700250,3700400,4029370,4029815,4031109

...

[I cut it off as the response is quite long]


usage patterns look similar to me. I See imageGet.sh also uses the sed command - perhaps it's a similar issue as before?


wget --help

GNU Wget 1.16.3, a non-interactive network retriever.

Usage: wget [OPTION]... [URL]...

Mandatory arguments to long options are mandatory for short options too.

Startup:

-V, --version display the version of Wget and exit

-h, --help print this help

-b, --background go to background after startup

-e, --execute=COMMAND execute a `.wgetrc'-style command

Logging and input file:

-o, --output-file=FILE log messages to FILE

-a, --append-output=FILE append messages to FILE

-d, --debug print lots of debugging information

-q, --quiet quiet (no output)

-v, --verbose be verbose (this is the default)

-nv, --no-verbose turn off verboseness, without being quiet

--report-speed=TYPE output bandwidth as TYPE. TYPE can be bits

-i, --input-file=FILE download URLs found in local or external FILE

-F, --force-html treat input file as HTML

-B, --base=URL resolves HTML input-file links (-i -F)

relative to URL

--config=FILE specify config file to use

--no-config do not read any config file

Download:

-t, --tries=NUMBER set number of retries to NUMBER (0 unlimits)

--retry-connrefused retry even if connection is refused

-O, --output-document=FILE write documents to FILE

-nc, --no-clobber skip downloads that would download to

existing files (overwriting them)

-c, --continue resume getting a partially-downloaded file

--start-pos=OFFSET start downloading from zero-based position OFFSET

--progress=TYPE select progress gauge type

--show-progress display the progress bar in any verbosity mode

-N, --timestamping don't re-retrieve files unless newer than

local

--no-use-server-timestamps don't set the local file's timestamp by

the one on the server

-S, --server-response print server response

--spider don't download anything

-T, --timeout=SECONDS set all timeout values to SECONDS

--dns-timeout=SECS set the DNS lookup timeout to SECS

--connect-timeout=SECS set the connect timeout to SECS

--read-timeout=SECS set the read timeout to SECS

-w, --wait=SECONDS wait SECONDS between retrievals

--waitretry=SECONDS wait 1..SECONDS between retries of a retrieval

--random-wait wait from 0.5*WAIT...1.5*WAIT secs between retrievals

--no-proxy explicitly turn off proxy

-Q, --quota=NUMBER set retrieval quota to NUMBER

--bind-address=ADDRESS bind to ADDRESS (hostname or IP) on local host

--limit-rate=RATE limit download rate to RATE

--no-dns-cache disable caching DNS lookups

--restrict-file-names=OS restrict chars in file names to ones OS allows

--ignore-case ignore case when matching files/directories

-4, --inet4-only connect only to IPv4 addresses

-6, --inet6-only connect only to IPv6 addresses

--prefer-family=FAMILY connect first to addresses of specified family,

one of IPv6, IPv4, or none

--user=USER set both ftp and http user to USER

--password=PASS set both ftp and http password to PASS

--ask-password prompt for passwords

--no-iri turn off IRI support

--local-encoding=ENC use ENC as the local encoding for IRIs

--remote-encoding=ENC use ENC as the default remote encoding

--unlink remove file before clobber

Directories:

-nd, --no-directories don't create directories

-x, --force-directories force creation of directories

-nH, --no-host-directories don't create host directories

--protocol-directories use protocol name in directories

-P, --directory-prefix=PREFIX save files to PREFIX/..

--cut-dirs=NUMBER ignore NUMBER remote directory components

HTTP options:

--http-user=USER set http user to USER

--http-password=PASS set http password to PASS

--no-cache disallow server-cached data

--default-page=NAME change the default page name (normally

this is 'index.html'.)

-E, --adjust-extension save HTML/CSS documents with proper extensions

--ignore-length ignore 'Content-Length' header field

--header=STRING insert STRING among the headers

--max-redirect maximum redirections allowed per page

--proxy-user=USER set USER as proxy username

--proxy-password=PASS set PASS as proxy password

--referer=URL include 'Referer: URL' header in HTTP request

--save-headers save the HTTP headers to file

-U, --user-agent=AGENT identify as AGENT instead of Wget/VERSION

--no-http-keep-alive disable HTTP keep-alive (persistent connections)

--no-cookies don't use cookies

--load-cookies=FILE load cookies from FILE before session

--save-cookies=FILE save cookies to FILE after session

--keep-session-cookies load and save session (non-permanent) cookies

--post-data=STRING use the POST method; send STRING as the data

--post-file=FILE use the POST method; send contents of FILE

--method=HTTPMethod use method "HTTPMethod" in the request

--body-data=STRING send STRING as data. --method MUST be set

--body-file=FILE send contents of FILE. --method MUST be set

--content-disposition honor the Content-Disposition header when

choosing local file names (EXPERIMENTAL)

--content-on-error output the received content on server errors

--auth-no-challenge send Basic HTTP authentication information

without first waiting for the server's

challenge

HTTPS (SSL/TLS) options:

--secure-protocol=PR choose secure protocol, one of auto, SSLv2,

SSLv3, TLSv1 and PFS

--https-only only follow secure HTTPS links

--no-check-certificate don't validate the server's certificate

--certificate=FILE client certificate file

--certificate-type=TYPE client certificate type, PEM or DER

--private-key=FILE private key file

--private-key-type=TYPE private key type, PEM or DER

--ca-certificate=FILE file with the bundle of CAs

--ca-directory=DIR directory where hash list of CAs is stored

--crl-file=FILE file with bundle of CRLs

--random-file=FILE file with random data for seeding the SSL PRNG

--egd-file=FILE file naming the EGD socket with random data

FTP options:

--ftp-user=USER set ftp user to USER

--ftp-password=PASS set ftp password to PASS

--no-remove-listing don't remove '.listing' files

--no-glob turn off FTP file name globbing

--no-passive-ftp disable the "passive" transfer mode

--preserve-permissions preserve remote file permissions

--retr-symlinks when recursing, get linked-to files (not dir)

WARC options:

--warc-file=FILENAME save request/response data to a .warc.gz file

--warc-header=STRING insert STRING into the warcinfo record

--warc-max-size=NUMBER set maximum size of WARC files to NUMBER

--warc-cdx write CDX index files

--warc-dedup=FILENAME do not store records listed in this CDX file

--no-warc-compression do not compress WARC files with GZIP

--no-warc-digests do not calculate SHA1 digests

--no-warc-keep-log do not store the log file in a WARC record

--warc-tempdir=DIRECTORY location for temporary files created by the

WARC writer

Recursive download:

-r, --recursive specify recursive download

-l, --level=NUMBER maximum recursion depth (inf or 0 for infinite)

--delete-after delete files locally after downloading them

-k, --convert-links make links in downloaded HTML or CSS point to

local files

--backups=N before writing file X, rotate up to N backup files

-K, --backup-converted before converting file X, back up as X.orig

-m, --mirror shortcut for -N -r -l inf --no-remove-listing

-p, --page-requisites get all images, etc. needed to display HTML page

--strict-comments turn on strict (SGML) handling of HTML comments

Recursive accept/reject:

-A, --accept=LIST comma-separated list of accepted extensions

-R, --reject=LIST comma-separated list of rejected extensions

--accept-regex=REGEX regex matching accepted URLs

--reject-regex=REGEX regex matching rejected URLs

--regex-type=TYPE regex type (posix)

-D, --domains=LIST comma-separated list of accepted domains

--exclude-domains=LIST comma-separated list of rejected domains

--follow-ftp follow FTP links from HTML documents

--follow-tags=LIST comma-separated list of followed HTML tags

--ignore-tags=LIST comma-separated list of ignored HTML tags

-H, --span-hosts go to foreign hosts when recursive

-L, --relative follow relative links only

-I, --include-directories=LIST list of allowed directories

--trust-server-names use the name specified by the redirection

URL's last component

-X, --exclude-directories=LIST list of excluded directories

-np, --no-parent don't ascend to the parent directory

I assume it's supposed to create the file in the same dir as imageGet.sh right?


$ /Users/Telecart/Downloads/Ubooquity/PageBuilder/imageGet.sh -c "action comics" -d 2016

Comic: action comics

Date: 2016

Downloading image

action%20comics


but there's no files created anywhere as far as I can tell...


tried it with -r, not giving different result...


$ /Users/Telecart/Downloads/Ubooquity/PageBuilder/imageGet.sh --comic "Action Comics" --date "2016" -r

Comic: Action Comics

Date: 2016

Result:

Downloading image

Action%20Comics




Btw, books.css I think got left behind at some point and now shows default theme look & feel. I just copied over it from old folder and it looks more in-line.

...which is also why page 2 for marvel looks different than page 1!

mostly works - files being generated with data, except can't get any images

No matter what I consistently get mv: rename /Path/To/Name of Comics 2016.jpg to /Path/To/Name of Comics (2016)/folder.jpg: No such file or directory

Done!


Any thoughts?

great! Will update when I get a chance to test it out!


Other Mac users may still need to install wget I assume (and to do that, need homebrew and OpenSSL)

Otherwise no reason this wouldn't work on a Mac