X-Git-Url: http://git.vpit.fr/?p=perl%2Fmodules%2FScope-Upper.git;a=blobdiff_plain;f=Upper.xs;h=efc799ab07f3baa44db5375b493261c05fbf4f49;hp=a002eac9a3c976727316ee64f6b8c7f344020bb2;hb=f0ee3a495f7f12754a9e79ef83f9affc69ad2f66;hpb=6e6aa503ec0f8dc6f7c01ac956f57d34eb1f7eda diff --git a/Upper.xs b/Upper.xs index a002eac..efc799a 100644 --- a/Upper.xs +++ b/Upper.xs @@ -6,21 +6,21 @@ #include "perl.h" #include "XSUB.h" -#define __PACKAGE__ "Scope::Upper" +/* --- XS helpers ---------------------------------------------------------- */ -#ifndef SU_DEBUG -# define SU_DEBUG 0 -#endif +#define XSH_PACKAGE "Scope::Upper" + +#include "xsh/caps.h" +#include "xsh/util.h" +#include "xsh/debug.h" /* --- Compatibility ------------------------------------------------------- */ -#ifndef NOOP -# define NOOP +/* perl 5.23.8 onwards has a revamped context system */ +#if XSH_HAS_PERL(5, 23, 8) +# define SU_HAS_NEW_CXT #endif -#ifndef dNOOP -# define dNOOP -#endif #ifndef dVAR # define dVAR dNOOP @@ -42,20 +42,6 @@ # define PERL_UNUSED_VAR(V) #endif -#ifndef STMT_START -# define STMT_START do -#endif - -#ifndef STMT_END -# define STMT_END while (0) -#endif - -#if SU_DEBUG -# define SU_D(X) STMT_START X STMT_END -#else -# define SU_D(X) -#endif - #ifndef Newx # define Newx(v, n, c) New(0, v, n, c) #endif @@ -217,48 +203,31 @@ static U8 su_op_gimme_reverse(U8 gimme) { # define NEGATIVE_INDICES_VAR "NEGATIVE_INDICES" #endif -#define SU_HAS_PERL(R, V, S) (PERL_REVISION > (R) || (PERL_REVISION == (R) && (PERL_VERSION > (V) || (PERL_VERSION == (V) && (PERL_SUBVERSION >= (S)))))) -#define SU_HAS_PERL_EXACT(R, V, S) ((PERL_REVISION == (R)) && (PERL_VERSION == (V)) && (PERL_SUBVERSION == (S))) - -/* --- Threads and multiplicity -------------------------------------------- */ - -#ifndef SU_MULTIPLICITY -# if defined(MULTIPLICITY) || defined(PERL_IMPLICIT_CONTEXT) -# define SU_MULTIPLICITY 1 -# else -# define SU_MULTIPLICITY 0 -# endif -#endif -#if SU_MULTIPLICITY && !defined(tTHX) -# define tTHX PerlInterpreter* -#endif - -#if SU_MULTIPLICITY && defined(USE_ITHREADS) && defined(dMY_CXT) && defined(MY_CXT) && defined(START_MY_CXT) && defined(MY_CXT_INIT) && (defined(MY_CXT_CLONE) || defined(dMY_CXT_SV)) -# define SU_THREADSAFE 1 -# ifndef MY_CXT_CLONE -# define MY_CXT_CLONE \ - dMY_CXT_SV; \ - my_cxt_t *my_cxtp = (my_cxt_t*)SvPVX(newSV(sizeof(my_cxt_t)-1)); \ - Copy(INT2PTR(my_cxt_t*, SvUV(my_cxt_sv)), my_cxtp, 1, my_cxt_t); \ - sv_setuv(my_cxt_sv, PTR2UV(my_cxtp)) -# endif +/* CX_ARGARRAY(cx): the AV at pad[0] of the CV associated with CXt_SUB + * context cx */ + +#if XSH_HAS_PERL(5, 23, 8) +# define CX_ARGARRAY(cx) \ + ((AV*)(AvARRAY(MUTABLE_AV( \ + PadlistARRAY(CvPADLIST(cx->blk_sub.cv))[ \ + CvDEPTH(cx->blk_sub.cv)]))[0])) +/* XXX is the new def ok to use in lvalue cxt? Formerly it assigned to + * blk_sub.argarray, now to pad[0]. Does this matter? + */ +# define CX_ARGARRAY_set(cx,ary) \ + (AvARRAY(MUTABLE_AV( \ + PadlistARRAY(CvPADLIST(cx->blk_sub.cv))[ \ + CvDEPTH(cx->blk_sub.cv)]))[0] = (SV*)(ary)) #else -# define SU_THREADSAFE 0 -# undef dMY_CXT -# define dMY_CXT dNOOP -# undef MY_CXT -# define MY_CXT su_globaldata -# undef START_MY_CXT -# define START_MY_CXT static my_cxt_t MY_CXT; -# undef MY_CXT_INIT -# define MY_CXT_INIT NOOP -# undef MY_CXT_CLONE -# define MY_CXT_CLONE NOOP +# define CX_ARGARRAY(cx) (cx->blk_sub.argarray) +# define CX_ARGARRAY_set(cx,ary) (cx->blk_sub.argarray = (ary)) #endif + /* --- Error messages ------------------------------------------------------ */ -static const char su_stack_smash[] = "Cannot target a scope outside of the current stack"; +static const char su_stack_smash[] = "Cannot target a scope outside of the current stack"; +static const char su_no_such_target[] = "No targetable %s scope in the current stack"; /* --- Unique context ID global storage ------------------------------------ */ @@ -275,22 +244,14 @@ static su_uv_array su_uid_seq_counter; static perl_mutex su_uid_seq_counter_mutex; -#define SU_LOCK(M) MUTEX_LOCK(M) -#define SU_UNLOCK(M) MUTEX_UNLOCK(M) - -#else /* USE_ITHREADS */ - -#define SU_LOCK(M) -#define SU_UNLOCK(M) - -#endif /* !USE_ITHREADS */ +#endif /* USE_ITHREADS */ static UV su_uid_seq_next(pTHX_ UV depth) { #define su_uid_seq_next(D) su_uid_seq_next(aTHX_ (D)) UV seq; UV *seqs; - SU_LOCK(&su_uid_seq_counter_mutex); + XSH_LOCK(&su_uid_seq_counter_mutex); seqs = su_uid_seq_counter.seqs; @@ -307,7 +268,7 @@ static UV su_uid_seq_next(pTHX_ UV depth) { seq = ++seqs[depth]; - SU_UNLOCK(&su_uid_seq_counter_mutex); + XSH_UNLOCK(&su_uid_seq_counter_mutex); return seq; } @@ -334,53 +295,47 @@ static UV su_uid_depth(pTHX_ I32 cxix) { } typedef struct { - su_uid **map; - STRLEN used; - STRLEN alloc; + su_uid *map; + STRLEN used; + STRLEN alloc; } su_uid_storage; static void su_uid_storage_dup(pTHX_ su_uid_storage *new_cxt, const su_uid_storage *old_cxt, UV max_depth) { #define su_uid_storage_dup(N, O, D) su_uid_storage_dup(aTHX_ (N), (O), (D)) - su_uid **old_map = old_cxt->map; + su_uid *old_map = old_cxt->map; if (old_map) { - su_uid **new_map = new_cxt->map; - STRLEN old_used = old_cxt->used; - STRLEN new_used, new_alloc; - STRLEN i; + su_uid *new_map = new_cxt->map; + STRLEN old_used = old_cxt->used; + STRLEN new_used, new_alloc; + STRLEN i; - new_used = max_depth < old_used ? max_depth : old_used; + new_used = max_depth < old_used ? max_depth : old_used; new_cxt->used = new_used; - if (new_used <= new_cxt->alloc) - new_alloc = new_cxt->alloc; - else { - new_alloc = new_used; - Renew(new_map, new_alloc, su_uid *); - for (i = new_cxt->alloc; i < new_alloc; ++i) - new_map[i] = NULL; + if (new_used <= new_cxt->alloc) { + new_alloc = new_cxt->alloc; + } else { + new_alloc = new_used; + Renew(new_map, new_alloc, su_uid); new_cxt->map = new_map; new_cxt->alloc = new_alloc; } for (i = 0; i < new_alloc; ++i) { - su_uid *new_uid = new_map[i]; + su_uid *new_uid = new_map + i; if (i < new_used) { /* => i < max_depth && i < old_used */ - su_uid *old_uid = old_map[i]; + su_uid *old_uid = old_map + i; if (old_uid && (old_uid->flags & SU_UID_ACTIVE)) { - if (!new_uid) { - Newx(new_uid, 1, su_uid); - new_map[i] = new_uid; - } *new_uid = *old_uid; continue; } } - if (new_uid) - new_uid->flags &= ~SU_UID_ACTIVE; + new_uid->seq = 0; + new_uid->flags = 0; } } @@ -409,7 +364,7 @@ typedef struct { /* --- uplevel() data tokens and global storage ---------------------------- */ -#define SU_UPLEVEL_HIJACKS_RUNOPS SU_HAS_PERL(5, 8, 0) +#define SU_UPLEVEL_HIJACKS_RUNOPS XSH_HAS_PERL(5, 8, 0) typedef struct { void *next; @@ -456,7 +411,7 @@ static su_uplevel_ud *su_uplevel_ud_new(pTHX) { si->si_stack = newAV(); AvREAL_off(si->si_stack); si->si_cxstack = NULL; - si->si_cxmax = 0; + si->si_cxmax = -1; sud->si = si; @@ -471,16 +426,7 @@ static void su_uplevel_ud_delete(pTHX_ su_uplevel_ud *sud) { SvREFCNT_dec(si->si_stack); Safefree(si); - if (sud->tmp_uid_storage.map) { - su_uid **map = sud->tmp_uid_storage.map; - STRLEN alloc = sud->tmp_uid_storage.alloc; - STRLEN i; - - for (i = 0; i < alloc; ++i) - Safefree(map[i]); - - Safefree(map); - } + Safefree(sud->tmp_uid_storage.map); Safefree(sud); @@ -499,21 +445,42 @@ typedef struct { /* --- Global data --------------------------------------------------------- */ -#define MY_CXT_KEY __PACKAGE__ "::_guts" XS_VERSION - typedef struct { char *stack_placeholder; su_unwind_storage unwind_storage; su_yield_storage yield_storage; su_uplevel_storage uplevel_storage; su_uid_storage uid_storage; -} my_cxt_t; +} xsh_user_cxt_t; + +#define XSH_THREADS_USER_CONTEXT 1 +#define XSH_THREADS_USER_CLONE_NEEDS_DUP 0 +#define XSH_THREADS_COMPILE_TIME_PROTECTION 0 + +#if XSH_THREADSAFE + +static void xsh_user_clone(pTHX_ const xsh_user_cxt_t *old_cxt, xsh_user_cxt_t *new_cxt) { + new_cxt->uplevel_storage.top = NULL; + new_cxt->uplevel_storage.root = NULL; + new_cxt->uplevel_storage.count = 0; + + new_cxt->uid_storage.map = NULL; + new_cxt->uid_storage.used = 0; + new_cxt->uid_storage.alloc = 0; + + su_uid_storage_dup(&new_cxt->uid_storage, &old_cxt->uid_storage, + old_cxt->uid_storage.used); + + return; +} -START_MY_CXT +#endif /* XSH_THREADSAFE */ + +#include "xsh/threads.h" /* --- Stack manipulations ------------------------------------------------- */ -#define SU_SAVE_PLACEHOLDER() save_pptr(&MY_CXT.stack_placeholder) +#define SU_SAVE_PLACEHOLDER() save_pptr(&XSH_CXT.stack_placeholder) #define SU_SAVE_DESTRUCTOR_SIZE 3 #define SU_SAVE_PLACEHOLDER_SIZE 3 @@ -544,11 +511,11 @@ START_MY_CXT #define SU_SAVE_GVCV_SIZE SU_SAVE_DESTRUCTOR_SIZE -#if !SU_HAS_PERL(5, 8, 9) +#if !XSH_HAS_PERL(5, 8, 9) # define SU_SAVE_GP_SIZE 6 -#elif !SU_HAS_PERL(5, 13, 0) || (SU_RELEASE && SU_HAS_PERL_EXACT(5, 13, 0)) +#elif !XSH_HAS_PERL(5, 13, 0) || (SU_RELEASE && XSH_HAS_PERL_EXACT(5, 13, 0)) # define SU_SAVE_GP_SIZE 3 -#elif !SU_HAS_PERL(5, 13, 8) +#elif !XSH_HAS_PERL(5, 13, 8) # define SU_SAVE_GP_SIZE 4 #else # define SU_SAVE_GP_SIZE 3 @@ -575,13 +542,14 @@ static I32 su_av_key2idx(pTHX_ AV *av, I32 key) { return key; /* Added by MJD in perl-5.8.1 with 6f12eb6d2a1dfaf441504d869b27d2e40ef4966a */ -#if SU_HAS_PERL(5, 8, 1) +#if XSH_HAS_PERL(5, 8, 1) if (SvRMAGICAL(av)) { const MAGIC * const tied_magic = mg_find((SV *) av, PERL_MAGIC_tied); if (tied_magic) { - SV * const * const negative_indices_glob = - hv_fetch(SvSTASH(SvRV(SvTIED_obj((SV *) (av), tied_magic))), - NEGATIVE_INDICES_VAR, 16, 0); + SV * const * const negative_indices_glob = hv_fetch( + SvSTASH(SvRV(SvTIED_obj((SV *) (av), tied_magic))), + NEGATIVE_INDICES_VAR, sizeof(NEGATIVE_INDICES_VAR)-1, 0 + ); if (negative_indices_glob && SvTRUE(GvSV(*negative_indices_glob))) return key; } @@ -695,7 +663,7 @@ static void su_save_helem(pTHX_ HV *hv, SV *keysv, SV *val) { /* ... Saving code slots from a glob ....................................... */ -#if !SU_HAS_PERL(5, 10, 0) && !defined(mro_method_changed_in) +#if !XSH_HAS_PERL(5, 10, 0) && !defined(mro_method_changed_in) # define mro_method_changed_in(G) PL_sub_generation++ #endif @@ -758,7 +726,7 @@ typedef struct { /* ... Reap ................................................................ */ -#define SU_SAVE_LAST_CX (!SU_HAS_PERL(5, 8, 4) || (SU_HAS_PERL(5, 9, 5) && !SU_HAS_PERL(5, 14, 0)) || SU_HAS_PERL(5, 15, 0)) +#define SU_SAVE_LAST_CX (!XSH_HAS_PERL(5, 8, 4) || (XSH_HAS_PERL(5, 9, 5) && !XSH_HAS_PERL(5, 14, 0)) || XSH_HAS_PERL(5, 15, 0)) typedef struct { su_ud_common ci; @@ -775,11 +743,8 @@ static void su_call(pTHX_ SV *cb) { dSP; - SU_D({ - PerlIO_printf(Perl_debug_log, - "@@@ call scope_ix=%2d save_ix=%2d\n", - PL_scopestack_ix, PL_savestack_ix); - }); + XSH_D(su_debug_log("@@@ call scope_ix=%2d save_ix=%2d\n", + PL_scopestack_ix, PL_savestack_ix)); ENTER; SAVETMPS; @@ -933,13 +898,12 @@ static void su_localize(pTHX_ void *ud_) { #endif } - SU_D({ + XSH_D({ SV *z = newSV(0); SvUPGRADE(z, t); - PerlIO_printf(Perl_debug_log, "%p: === localize a %s\n",ud, sv_reftype(z, 0)); - PerlIO_printf(Perl_debug_log, - "%p: depth=%2d scope_ix=%2d save_ix=%2d\n", - ud, SU_UD_DEPTH(ud), PL_scopestack_ix, PL_savestack_ix); + su_debug_log("%p: === localize a %s\n",ud, sv_reftype(z, 0)); + su_debug_log("%p: depth=%2d scope_ix=%2d save_ix=%2d\n", + ud, SU_UD_DEPTH(ud), PL_scopestack_ix, PL_savestack_ix); SvREFCNT_dec(z); }); @@ -978,17 +942,21 @@ static void su_localize(pTHX_ void *ud_) { /* ... Unique context ID ................................................... */ +/* We must pass the index because XSH_CXT.uid_storage might be reallocated + * between the UID fetch and the invalidation at the end of scope. */ + typedef struct { su_ud_common ci; - su_uid *uid; + I32 idx; } su_ud_uid; -#define SU_UD_UID_UID(U) (((su_ud_uid *) (U))->uid) - static void su_uid_drop(pTHX_ void *ud_) { - su_uid *uid = ud_; + su_ud_uid *ud = ud_; + dXSH_CXT; + + XSH_CXT.uid_storage.map[ud->idx].flags &= ~SU_UID_ACTIVE; - uid->flags &= ~SU_UID_ACTIVE; + SU_UD_FREE(ud); return; } @@ -998,7 +966,7 @@ static void su_uid_drop(pTHX_ void *ud_) { #ifdef DEBUGGING # define SU_CXNAME(C) PL_block_type[CxTYPE(C)] #else -# if SU_HAS_PERL(5, 11, 0) +# if XSH_HAS_PERL(5, 11, 0) static const char *su_block_type[] = { "NULL", "WHEN", @@ -1013,7 +981,7 @@ static const char *su_block_type[] = { "EVAL", "SUBST" }; -# elif SU_HAS_PERL(5, 9, 3) +# elif XSH_HAS_PERL(5, 9, 3) static const char *su_block_type[] = { "NULL", "SUB", @@ -1046,31 +1014,29 @@ static void su_pop(pTHX_ void *ud) { I32 depth, base, mark, *origin; depth = SU_UD_DEPTH(ud); - SU_D( - PerlIO_printf(Perl_debug_log, - "%p: --- pop a %s\n" - "%p: leave scope at depth=%2d scope_ix=%2d cur_top=%2d cur_base=%2d\n", - ud, SU_CXNAME(cxstack + cxstack_ix), - ud, depth, PL_scopestack_ix,PL_savestack_ix,PL_scopestack[PL_scopestack_ix]) - ); + XSH_D(su_debug_log( + "%p: --- pop a %s\n" + "%p: leave scope at depth=%2d scope_ix=%2d cur_top=%2d cur_base=%2d\n", + ud, SU_CXNAME(cxstack + cxstack_ix), + ud, depth, PL_scopestack_ix,PL_savestack_ix,PL_scopestack[PL_scopestack_ix] + )); origin = SU_UD_ORIGIN(ud); mark = origin[depth]; base = origin[depth - 1]; - SU_D(PerlIO_printf(Perl_debug_log, - "%p: original scope was %*c top=%2d base=%2d\n", + XSH_D(su_debug_log("%p: original scope was %*c top=%2d base=%2d\n", ud, 24, ' ', mark, base)); if (base < mark) { -#if SU_HAS_PERL(5, 19, 4) +#if XSH_HAS_PERL(5, 19, 4) I32 save = -1; PERL_CONTEXT *cx; #endif - SU_D(PerlIO_printf(Perl_debug_log, "%p: clear leftovers\n", ud)); + XSH_D(su_debug_log("%p: clear leftovers\n", ud)); -#if SU_HAS_PERL(5, 19, 4) +#if XSH_HAS_PERL(5, 19, 4) cx = cxstack + cxstack_ix; if (CxTYPE(cx) == CXt_SUB || CxTYPE(cx) == CXt_FORMAT) save = PL_scopestack[cx->blk_oldscopesp - 1]; @@ -1079,7 +1045,7 @@ static void su_pop(pTHX_ void *ud) { PL_savestack_ix = mark; leave_scope(base); -#if SU_HAS_PERL(5, 19, 4) +#if XSH_HAS_PERL(5, 19, 4) if (CxTYPE(cx) == CXt_SUB || CxTYPE(cx) == CXt_FORMAT) PL_scopestack[cx->blk_oldscopesp - 1] = save; #endif @@ -1092,27 +1058,24 @@ static void su_pop(pTHX_ void *ud) { U8 pad; if ((pad = SU_UD_PAD(ud)) > 0) { - dMY_CXT; + dXSH_CXT; do { - SU_D(PerlIO_printf(Perl_debug_log, + XSH_D(su_debug_log( "%p: push a pad slot at depth=%2d scope_ix=%2d save_ix=%2d\n", ud, depth, PL_scopestack_ix, PL_savestack_ix)); SU_SAVE_PLACEHOLDER(); } while (--pad); } - SU_D(PerlIO_printf(Perl_debug_log, - "%p: push destructor at depth=%2d scope_ix=%2d save_ix=%2d\n", - ud, depth, PL_scopestack_ix, PL_savestack_ix)); + XSH_D(su_debug_log( + "%p: push destructor at depth=%2d scope_ix=%2d save_ix=%2d\n", + ud, depth, PL_scopestack_ix, PL_savestack_ix)); SAVEDESTRUCTOR_X(su_pop, ud); } else { switch (SU_UD_TYPE(ud)) { case SU_UD_TYPE_REAP: { - SU_D({ - PerlIO_printf(Perl_debug_log, - "%p: === reap\n%p: depth=%2d scope_ix=%2d save_ix=%2d\n", - ud, ud, SU_UD_DEPTH(ud), PL_scopestack_ix, PL_savestack_ix); - }); + XSH_D(su_debug_log("%p: === reap\n%p: depth=%2d scope_ix=%2d save_ix=%2d\n", + ud, ud, SU_UD_DEPTH(ud), PL_scopestack_ix, PL_savestack_ix)); SAVEDESTRUCTOR_X(su_call, SU_UD_REAP_CB(ud)); SU_UD_FREE(ud); break; @@ -1122,14 +1085,12 @@ static void su_pop(pTHX_ void *ud) { SU_UD_LOCALIZE_FREE(ud); break; case SU_UD_TYPE_UID: - SAVEDESTRUCTOR_X(su_uid_drop, SU_UD_UID_UID(ud)); - SU_UD_FREE(ud); + SAVEDESTRUCTOR_X(su_uid_drop, ud); break; } } - SU_D(PerlIO_printf(Perl_debug_log, - "%p: --- end pop: cur_top=%2d == cur_base=%2d\n", + XSH_D(su_debug_log("%p: --- end pop: cur_top=%2d == cur_base=%2d\n", ud, PL_savestack_ix, PL_scopestack[PL_scopestack_ix])); } @@ -1140,23 +1101,42 @@ static I32 su_init(pTHX_ void *ud, I32 cxix, I32 size) { I32 i, depth, offset, base, *origin; U8 pad; - SU_D(PerlIO_printf(Perl_debug_log, "%p: ### init for cx %d\n", ud, cxix)); - - if (size <= SU_SAVE_DESTRUCTOR_SIZE) + XSH_D(su_debug_log("%p: ### init for cx %d\n", ud, cxix)); + + /* su_pop() is going to be called from leave_scope(), so before pushing the + * next callback, we'll want to flush the current scope stack slice first. + * However, if we want the next callback not to be processed immediately by + * the current leave_scope(), we'll need to hide it by artificially + * incrementing the scope stack marker before. For the intermediate bumps, + * we will only need a bump of SU_SAVE_DESTRUCTOR_SIZE items, but for the + * last one we will need a bump of size items. However, in order to preserve + * the natural ordering between scope stack markers, we cannot bump lower + * markers more than higher ones. This is why we bump the intermediate markers + * by the smallest multiple of SU_SAVE_PLACEHOLDER_SIZE greater or equal to + * max(SU_SAVE_DESTRUCTOR_SIZE, size). */ + + if (size <= SU_SAVE_DESTRUCTOR_SIZE) { pad = 0; - else { + } else { I32 extra = size - SU_SAVE_DESTRUCTOR_SIZE; pad = extra / SU_SAVE_PLACEHOLDER_SIZE; if (extra % SU_SAVE_PLACEHOLDER_SIZE) ++pad; } offset = SU_SAVE_DESTRUCTOR_SIZE + SU_SAVE_PLACEHOLDER_SIZE * pad; - - SU_D(PerlIO_printf(Perl_debug_log, "%p: size=%d pad=%d offset=%d\n", - ud, size, pad, offset)); + XSH_D(su_debug_log("%p: size=%d pad=%d offset=%d\n", ud, size, pad, offset)); depth = PL_scopestack_ix - cxstack[cxix].blk_oldscopesp; - SU_D(PerlIO_printf(Perl_debug_log, "%p: going down to depth %d\n", ud, depth)); + XSH_D(su_debug_log("%p: going down to depth %d\n", ud, depth)); + + /* We need to bump all the intermediary stack markers just in case an + * exception is thrown before the target scope is reached. Indeed, in this + * case there might be arbitrary many scope frames flushed at the same time, + * and since we cannot know in advance whether this will happen or not, we + * have to make sure the final frame is protected for the actual action. But + * of course, in order to do that, we also need to bump all the previous stack + * markers. If not for this, it should have been possible to just bump the two + * next frames in su_pop(). */ Newx(origin, depth + 1, I32); base = PL_scopestack_ix - depth; @@ -1164,6 +1144,7 @@ static I32 su_init(pTHX_ void *ud, I32 cxix, I32 size) { PL_scopestack[base] += size; for (i = 1; i < depth; ++i) { I32 j = i + base; + /* origin[depth - i] == PL_scopestack[PL_scopestack_ix - i] */ origin[i] = PL_scopestack[j]; PL_scopestack[j] += offset; } @@ -1177,26 +1158,23 @@ static I32 su_init(pTHX_ void *ud, I32 cxix, I32 size) { * stack. */ if (PL_savestack_ix + SU_SAVE_DESTRUCTOR_SIZE <= PL_scopestack[PL_scopestack_ix - 1]) { - dMY_CXT; + dXSH_CXT; do { - SU_D(PerlIO_printf(Perl_debug_log, - "%p: push a fake slot at scope_ix=%2d save_ix=%2d\n", - ud, PL_scopestack_ix, PL_savestack_ix)); + XSH_D(su_debug_log("%p: push a fake slot at scope_ix=%2d save_ix=%2d\n", + ud, PL_scopestack_ix, PL_savestack_ix)); SU_SAVE_PLACEHOLDER(); } while (PL_savestack_ix + SU_SAVE_DESTRUCTOR_SIZE <= PL_scopestack[PL_scopestack_ix - 1]); } - SU_D(PerlIO_printf(Perl_debug_log, - "%p: push first destructor at scope_ix=%2d save_ix=%2d\n", - ud, PL_scopestack_ix, PL_savestack_ix)); + XSH_D(su_debug_log("%p: push first destructor at scope_ix=%2d save_ix=%2d\n", + ud, PL_scopestack_ix, PL_savestack_ix)); SAVEDESTRUCTOR_X(su_pop, ud); - SU_D({ + XSH_D({ for (i = 0; i <= depth; ++i) { I32 j = PL_scopestack_ix - i; - PerlIO_printf(Perl_debug_log, - "%p: depth=%2d scope_ix=%2d saved_floor=%2d new_floor=%2d\n", - ud, i, j, origin[depth - i], + su_debug_log("%p: depth=%2d scope_ix=%2d saved_floor=%2d new_floor=%2d\n", + ud, i, j, origin[depth - i], i == 0 ? PL_savestack_ix : PL_scopestack[j]); } }); @@ -1207,15 +1185,15 @@ static I32 su_init(pTHX_ void *ud, I32 cxix, I32 size) { /* --- Unwind stack -------------------------------------------------------- */ static void su_unwind(pTHX_ void *ud_) { - dMY_CXT; - I32 cxix = MY_CXT.unwind_storage.cxix; - I32 items = MY_CXT.unwind_storage.items; + dXSH_CXT; + I32 cxix = XSH_CXT.unwind_storage.cxix; + I32 items = XSH_CXT.unwind_storage.items; I32 mark; PERL_UNUSED_VAR(ud_); - PL_stack_sp = MY_CXT.unwind_storage.savesp; -#if SU_HAS_PERL(5, 19, 4) + PL_stack_sp = XSH_CXT.unwind_storage.savesp; +#if XSH_HAS_PERL(5, 19, 4) { I32 i; SV **sp = PL_stack_sp; @@ -1231,27 +1209,26 @@ static void su_unwind(pTHX_ void *ud_) { mark = PL_markstack[cxstack[cxix].blk_oldmarksp]; *PL_markstack_ptr = PL_stack_sp - PL_stack_base - items; - SU_D({ + XSH_D({ I32 gimme = GIMME_V; - PerlIO_printf(Perl_debug_log, - "%p: cx=%d gimme=%s items=%d sp=%d oldmark=%d mark=%d\n", - &MY_CXT, cxix, + su_debug_log("%p: cx=%d gimme=%s items=%d sp=%d oldmark=%d mark=%d\n", + &XSH_CXT, cxix, gimme == G_VOID ? "void" : gimme == G_ARRAY ? "list" : "scalar", items, PL_stack_sp - PL_stack_base, *PL_markstack_ptr, mark); }); - PL_op = (OP *) &(MY_CXT.unwind_storage.return_op); + PL_op = (OP *) &(XSH_CXT.unwind_storage.return_op); PL_op = PL_op->op_ppaddr(aTHX); *PL_markstack_ptr = mark; - MY_CXT.unwind_storage.proxy_op.op_next = PL_op; - PL_op = &(MY_CXT.unwind_storage.proxy_op); + XSH_CXT.unwind_storage.proxy_op.op_next = PL_op; + PL_op = &(XSH_CXT.unwind_storage.proxy_op); } /* --- Yield --------------------------------------------------------------- */ -#if SU_HAS_PERL(5, 10, 0) +#if XSH_HAS_PERL(5, 10, 0) # define SU_RETOP_SUB(C) ((C)->blk_sub.retop) # define SU_RETOP_EVAL(C) ((C)->blk_eval.retop) # define SU_RETOP_LOOP(C) ((C)->blk_loop.my_op->op_lastop->op_next) @@ -1263,11 +1240,11 @@ static void su_unwind(pTHX_ void *ud_) { #endif static void su_yield(pTHX_ void *ud_) { - dMY_CXT; + dXSH_CXT; PERL_CONTEXT *cx; const char *which = ud_; - I32 cxix = MY_CXT.yield_storage.cxix; - I32 items = MY_CXT.yield_storage.items; + I32 cxix = XSH_CXT.yield_storage.cxix; + I32 items = XSH_CXT.yield_storage.items; opcode type = OP_NULL; U8 flags = 0; OP *next; @@ -1281,7 +1258,7 @@ static void su_yield(pTHX_ void *ud_) { OP *o = NULL; /* Is this actually a given/when block? This may occur only when yield was * called with HERE (or nothing) as the context. */ -#if SU_HAS_PERL(5, 10, 0) +#if XSH_HAS_PERL(5, 10, 0) if (cxix > 0) { PERL_CONTEXT *prev = cx - 1; U8 prev_type = CxTYPE(prev); @@ -1320,8 +1297,13 @@ static void su_yield(pTHX_ void *ud_) { case CXt_EVAL: o = SU_RETOP_EVAL(cx2); break; -#if SU_HAS_PERL(5, 11, 0) +#if XSH_HAS_PERL(5, 11, 0) +# if XSH_HAS_PERL(5, 23, 8) + case CXt_LOOP_ARY: + case CXt_LOOP_LIST: +# else case CXt_LOOP_FOR: +# endif case CXt_LOOP_PLAIN: case CXt_LOOP_LAZYSV: case CXt_LOOP_LAZYIV: @@ -1361,8 +1343,13 @@ static void su_yield(pTHX_ void *ud_) { type = CxTRYBLOCK(cx) ? OP_LEAVETRY : OP_LEAVEEVAL; next = SU_RETOP_EVAL(cx); break; -#if SU_HAS_PERL(5, 11, 0) +#if XSH_HAS_PERL(5, 11, 0) +# if XSH_HAS_PERL(5, 23, 8) + case CXt_LOOP_ARY: + case CXt_LOOP_LIST: +# else case CXt_LOOP_FOR: +# endif case CXt_LOOP_PLAIN: case CXt_LOOP_LAZYSV: case CXt_LOOP_LAZYIV: @@ -1372,7 +1359,7 @@ static void su_yield(pTHX_ void *ud_) { type = OP_LEAVELOOP; next = SU_RETOP_LOOP(cx); break; -#if SU_HAS_PERL(5, 10, 0) +#if XSH_HAS_PERL(5, 10, 0) case CXt_GIVEN: cxt_given: type = OP_LEAVEGIVEN; @@ -1380,7 +1367,7 @@ cxt_given: break; case CXt_WHEN: cxt_when: -#if SU_HAS_PERL(5, 15, 1) +#if XSH_HAS_PERL(5, 15, 1) type = OP_LEAVEWHEN; #else type = OP_BREAK; @@ -1398,8 +1385,8 @@ cxt_when: break; } - PL_stack_sp = MY_CXT.yield_storage.savesp; -#if SU_HAS_PERL(5, 19, 4) + PL_stack_sp = XSH_CXT.yield_storage.savesp; +#if XSH_HAS_PERL(5, 19, 4) { I32 i; SV **sp = PL_stack_sp; @@ -1420,16 +1407,16 @@ cxt_when: flags |= OP_GIMME_REVERSE(cx->blk_gimme); - MY_CXT.yield_storage.leave_op.op_type = type; - MY_CXT.yield_storage.leave_op.op_ppaddr = PL_ppaddr[type]; - MY_CXT.yield_storage.leave_op.op_flags = flags; - MY_CXT.yield_storage.leave_op.op_next = next; + XSH_CXT.yield_storage.leave_op.op_type = type; + XSH_CXT.yield_storage.leave_op.op_ppaddr = PL_ppaddr[type]; + XSH_CXT.yield_storage.leave_op.op_flags = flags; + XSH_CXT.yield_storage.leave_op.op_next = next; - PL_op = (OP *) &(MY_CXT.yield_storage.leave_op); + PL_op = (OP *) &(XSH_CXT.yield_storage.leave_op); PL_op = PL_op->op_ppaddr(aTHX); - MY_CXT.yield_storage.proxy_op.op_next = PL_op; - PL_op = &(MY_CXT.yield_storage.proxy_op); + XSH_CXT.yield_storage.proxy_op.op_next = PL_op; + PL_op = &(XSH_CXT.yield_storage.proxy_op); } /* --- Uplevel ------------------------------------------------------------- */ @@ -1441,53 +1428,51 @@ static su_uplevel_ud *su_uplevel_storage_new(pTHX_ I32 cxix) { #define su_uplevel_storage_new(I) su_uplevel_storage_new(aTHX_ (I)) su_uplevel_ud *sud; UV depth; - dMY_CXT; + dXSH_CXT; - sud = MY_CXT.uplevel_storage.root; + sud = XSH_CXT.uplevel_storage.root; if (sud) { - MY_CXT.uplevel_storage.root = sud->next; - MY_CXT.uplevel_storage.count--; + XSH_CXT.uplevel_storage.root = sud->next; + XSH_CXT.uplevel_storage.count--; } else { sud = su_uplevel_ud_new(); } - sud->next = MY_CXT.uplevel_storage.top; - MY_CXT.uplevel_storage.top = sud; + sud->next = XSH_CXT.uplevel_storage.top; + XSH_CXT.uplevel_storage.top = sud; depth = su_uid_depth(cxix); - su_uid_storage_dup(&sud->tmp_uid_storage, &MY_CXT.uid_storage, depth); - sud->old_uid_storage = MY_CXT.uid_storage; - MY_CXT.uid_storage = sud->tmp_uid_storage; + su_uid_storage_dup(&sud->tmp_uid_storage, &XSH_CXT.uid_storage, depth); + sud->old_uid_storage = XSH_CXT.uid_storage; + XSH_CXT.uid_storage = sud->tmp_uid_storage; return sud; } -#if SU_HAS_PERL(5, 13, 7) +#if XSH_HAS_PERL(5, 13, 7) static void su_uplevel_storage_delete(pTHX_ su_uplevel_ud *sud) { #define su_uplevel_storage_delete(S) su_uplevel_storage_delete(aTHX_ (S)) - dMY_CXT; + dXSH_CXT; - sud->tmp_uid_storage = MY_CXT.uid_storage; - MY_CXT.uid_storage = sud->old_uid_storage; + sud->tmp_uid_storage = XSH_CXT.uid_storage; + XSH_CXT.uid_storage = sud->old_uid_storage; { - su_uid **map; - UV i, alloc; + su_uid *map; + STRLEN i, alloc; map = sud->tmp_uid_storage.map; alloc = sud->tmp_uid_storage.alloc; - for (i = 0; i < alloc; ++i) { - if (map[i]) - map[i]->flags &= SU_UID_ACTIVE; - } + for (i = 0; i < alloc; ++i) + map[i].flags &= ~SU_UID_ACTIVE; } - MY_CXT.uplevel_storage.top = sud->next; + XSH_CXT.uplevel_storage.top = sud->next; - if (MY_CXT.uplevel_storage.count >= SU_UPLEVEL_STORAGE_SIZE) { + if (XSH_CXT.uplevel_storage.count >= SU_UPLEVEL_STORAGE_SIZE) { su_uplevel_ud_delete(sud); } else { - sud->next = MY_CXT.uplevel_storage.root; - MY_CXT.uplevel_storage.root = sud; - MY_CXT.uplevel_storage.count++; + sud->next = XSH_CXT.uplevel_storage.root; + XSH_CXT.uplevel_storage.root = sud; + XSH_CXT.uplevel_storage.count++; } } @@ -1535,7 +1520,7 @@ static int su_uplevel_goto_runops(pTHX) { switch (CxTYPE(cx)) { case CXt_SUB: if (CxHASARGS(cx)) { - argarray = cx->blk_sub.argarray; + argarray = CX_ARGARRAY(cx); goto done; } break; @@ -1549,9 +1534,9 @@ static int su_uplevel_goto_runops(pTHX) { done: if (argarray) { - dMY_CXT; + dXSH_CXT; - if (MY_CXT.uplevel_storage.top->cxix == cxix) { + if (XSH_CXT.uplevel_storage.top->cxix == cxix) { AV *args = GvAV(PL_defgv); I32 items = AvFILLp(args); @@ -1564,7 +1549,7 @@ done: PL_op = op = op->op_ppaddr(aTHX); -#if !SU_HAS_PERL(5, 13, 0) +#if !XSH_HAS_PERL(5, 13, 0) PERL_ASYNC_CHECK(); #endif } while (op); @@ -1610,8 +1595,8 @@ static void su_uplevel_restore(pTHX_ void *sus_) { * reached without a goto() happening, and the old argarray member is * actually our fake argarray. Destroy it properly in that case. */ if (cx->blk_sub.cv == sud->renamed) { - SvREFCNT_dec(cx->blk_sub.argarray); - cx->blk_sub.argarray = argarray; + SvREFCNT_dec(CX_ARGARRAY(cx)); + CX_ARGARRAY_set(cx, argarray); } CvDEPTH(sud->callback)--; @@ -1684,7 +1669,7 @@ found_it: CvDEPTH(target) = sud->target_depth - levels; PL_curstackinfo->si_cxix = i - 1; -#if !SU_HAS_PERL(5, 13, 1) +#if !XSH_HAS_PERL(5, 13, 1) /* Since $@ was maybe localized between the target frame and the uplevel * call, we forcefully flush the save stack to get rid of it and then * reset $@ to its proper value. Note that the the call to @@ -1718,22 +1703,22 @@ found_it: * pointer to the current context frame across this call. This means that we * can't free the temporary context stack we used for the uplevel call right * now, or that pointer upwards would point to garbage. */ -#if SU_HAS_PERL(5, 13, 7) +#if XSH_HAS_PERL(5, 13, 7) /* This issue has been fixed in perl with commit 8f89e5a9, which was made * public in perl 5.13.7. */ su_uplevel_storage_delete(sud); #else /* Otherwise, we just enqueue it back in the global storage list. */ { - dMY_CXT; + dXSH_CXT; - sud->tmp_uid_storage = MY_CXT.uid_storage; - MY_CXT.uid_storage = sud->old_uid_storage; + sud->tmp_uid_storage = XSH_CXT.uid_storage; + XSH_CXT.uid_storage = sud->old_uid_storage; - MY_CXT.uplevel_storage.top = sud->next; - sud->next = MY_CXT.uplevel_storage.root; - MY_CXT.uplevel_storage.root = sud; - MY_CXT.uplevel_storage.count++; + XSH_CXT.uplevel_storage.top = sud->next; + sud->next = XSH_CXT.uplevel_storage.root; + XSH_CXT.uplevel_storage.root = sud; + XSH_CXT.uplevel_storage.count++; } #endif @@ -1759,14 +1744,14 @@ static CV *su_cv_clone(pTHX_ CV *proto, GV *gv) { #endif CvGV_set(cv, gv); -#if SU_RELEASE && SU_HAS_PERL_EXACT(5, 21, 4) +#if SU_RELEASE && XSH_HAS_PERL_EXACT(5, 21, 4) CvNAMED_off(cv); #endif CvSTASH_set(cv, CvSTASH(proto)); /* Commit 4c74a7df, publicized with perl 5.13.3, began to add backrefs to * stashes. CvSTASH_set() started to do it as well with commit c68d95645 * (which was part of perl 5.13.7). */ -#if SU_HAS_PERL(5, 13, 3) && !SU_HAS_PERL(5, 13, 7) +#if XSH_HAS_PERL(5, 13, 3) && !XSH_HAS_PERL(5, 13, 7) if (CvSTASH(proto)) Perl_sv_add_backref(aTHX_ CvSTASH(proto), MUTABLE_SV(cv)); #endif @@ -1915,6 +1900,7 @@ static I32 su_uplevel(pTHX_ CV *callback, I32 cxix, I32 args) { if ((PL_op = PL_ppaddr[OP_ENTERSUB](aTHX))) { PERL_CONTEXT *sub_cx = cxstack + cxstack_ix; + AV *argarray = CX_ARGARRAY(cx); /* If pp_entersub() returns a non-null OP, it means that the callback is not * an XSUB. */ @@ -1922,7 +1908,7 @@ static I32 su_uplevel(pTHX_ CV *callback, I32 cxix, I32 args) { sud->callback = MUTABLE_CV(SvREFCNT_inc(callback)); CvDEPTH(callback)++; - if (CxHASARGS(cx) && cx->blk_sub.argarray) { + if (CxHASARGS(cx) && argarray) { /* The call to pp_entersub() has saved the current @_ (in XS terms, * GvAV(PL_defgv)) in the savearray member, and has created a new argarray * with what we put on the stack. But we want to fake up the same arguments @@ -1931,12 +1917,12 @@ static I32 su_uplevel(pTHX_ CV *callback, I32 cxix, I32 args) { AV *av = newAV(); AvREAL_off(av); AvREIFY_on(av); - av_extend(av, AvMAX(cx->blk_sub.argarray)); - AvFILLp(av) = AvFILLp(cx->blk_sub.argarray); - Copy(AvARRAY(cx->blk_sub.argarray), AvARRAY(av), AvFILLp(av) + 1, SV *); - sub_cx->blk_sub.argarray = av; + av_extend(av, AvMAX(argarray)); + AvFILLp(av) = AvFILLp(argarray); + Copy(AvARRAY(argarray), AvARRAY(av), AvFILLp(av) + 1, SV *); + CX_ARGARRAY_set(sub_cx, av); } else { - SvREFCNT_inc_simple_void(sub_cx->blk_sub.argarray); + SvREFCNT_inc_simple_void(CX_ARGARRAY(sub_cx)); } if (su_uplevel_goto_static(CvROOT(renamed))) { @@ -1984,57 +1970,50 @@ static I32 su_uplevel(pTHX_ CV *callback, I32 cxix, I32 args) { static su_uid *su_uid_storage_fetch(pTHX_ UV depth) { #define su_uid_storage_fetch(D) su_uid_storage_fetch(aTHX_ (D)) - su_uid **map, *uid; - STRLEN alloc; - dMY_CXT; + su_uid *map; + STRLEN alloc; + dXSH_CXT; - map = MY_CXT.uid_storage.map; - alloc = MY_CXT.uid_storage.alloc; + map = XSH_CXT.uid_storage.map; + alloc = XSH_CXT.uid_storage.alloc; if (depth >= alloc) { STRLEN i; - Renew(map, depth + 1, su_uid *); - for (i = alloc; i <= depth; ++i) - map[i] = NULL; - - MY_CXT.uid_storage.map = map; - MY_CXT.uid_storage.alloc = depth + 1; - } - - uid = map[depth]; + Renew(map, depth + 1, su_uid); + for (i = alloc; i <= depth; ++i) { + map[i].seq = 0; + map[i].flags = 0; + } - if (!uid) { - Newx(uid, 1, su_uid); - uid->seq = 0; - uid->flags = 0; - map[depth] = uid; + XSH_CXT.uid_storage.map = map; + XSH_CXT.uid_storage.alloc = depth + 1; } - if (depth >= MY_CXT.uid_storage.used) - MY_CXT.uid_storage.used = depth + 1; + if (depth >= XSH_CXT.uid_storage.used) + XSH_CXT.uid_storage.used = depth + 1; - return uid; + return map + depth; } static int su_uid_storage_check(pTHX_ UV depth, UV seq) { #define su_uid_storage_check(D, S) su_uid_storage_check(aTHX_ (D), (S)) su_uid *uid; - dMY_CXT; + dXSH_CXT; - if (depth >= MY_CXT.uid_storage.used) + if (depth >= XSH_CXT.uid_storage.used) return 0; - uid = MY_CXT.uid_storage.map[depth]; + uid = XSH_CXT.uid_storage.map + depth; - return uid && (uid->seq == seq) && (uid->flags & SU_UID_ACTIVE); + return (uid->seq == seq) && (uid->flags & SU_UID_ACTIVE); } static SV *su_uid_get(pTHX_ I32 cxix) { #define su_uid_get(I) su_uid_get(aTHX_ (I)) su_uid *uid; - SV *uid_sv; - UV depth; + SV *uid_sv; + UV depth; depth = su_uid_depth(cxix); uid = su_uid_storage_fetch(depth); @@ -2047,12 +2026,13 @@ static SV *su_uid_get(pTHX_ I32 cxix) { Newx(ud, 1, su_ud_uid); SU_UD_TYPE(ud) = SU_UD_TYPE_UID; - ud->uid = uid; + ud->idx = depth; su_init(ud, cxix, SU_SAVE_DESTRUCTOR_SIZE); } uid_sv = sv_newmortal(); sv_setpvf(uid_sv, "%"UVuf"-%"UVuf, depth, uid->seq); + return uid_sv; } @@ -2128,7 +2108,7 @@ static I32 su_context_skip_db(pTHX_ I32 cxix) { PERL_CONTEXT *cx = cxstack + i; switch (CxTYPE(cx)) { -#if SU_HAS_PERL(5, 17, 1) +#if XSH_HAS_PERL(5, 17, 1) case CXt_LOOP_PLAIN: #endif case CXt_BLOCK: @@ -2164,11 +2144,11 @@ static I32 su_context_normalize_up(pTHX_ I32 cxix) { PERL_CONTEXT *prev = cx - 1; switch (CxTYPE(prev)) { -#if SU_HAS_PERL(5, 10, 0) +#if XSH_HAS_PERL(5, 10, 0) case CXt_GIVEN: case CXt_WHEN: #endif -#if SU_HAS_PERL(5, 11, 0) +#if XSH_HAS_PERL(5, 11, 0) /* That's the only subcategory that can cause an extra BLOCK context */ case CXt_LOOP_PLAIN: #else @@ -2200,11 +2180,11 @@ static I32 su_context_normalize_down(pTHX_ I32 cxix) { PERL_CONTEXT *cx = next - 1; switch (CxTYPE(cx)) { -#if SU_HAS_PERL(5, 10, 0) +#if XSH_HAS_PERL(5, 10, 0) case CXt_GIVEN: case CXt_WHEN: #endif -#if SU_HAS_PERL(5, 11, 0) +#if XSH_HAS_PERL(5, 11, 0) /* That's the only subcategory that can cause an extra BLOCK context */ case CXt_LOOP_PLAIN: #else @@ -2235,8 +2215,13 @@ static I32 su_context_gimme(pTHX_ I32 cxix) { switch (CxTYPE(cx)) { /* gimme is always G_ARRAY for loop contexts. */ -#if SU_HAS_PERL(5, 11, 0) +#if XSH_HAS_PERL(5, 11, 0) +# if XSH_HAS_PERL(5, 23, 8) + case CXt_LOOP_ARY: + case CXt_LOOP_LIST: +# else case CXt_LOOP_FOR: +# endif case CXt_LOOP_PLAIN: case CXt_LOOP_LAZYSV: case CXt_LOOP_LAZYIV: @@ -2268,90 +2253,62 @@ static I32 su_context_gimme(pTHX_ I32 cxix) { return G_VOID; } -/* --- Global setup/teardown ----------------------------------------------- */ - -static VOL U32 su_initialized = 0; +/* --- Module setup/teardown ----------------------------------------------- */ -static void su_global_teardown(pTHX_ void *root) { - if (!su_initialized) - return; +static void xsh_user_global_setup(pTHX) { + HV *stash; -#if SU_MULTIPLICITY - if (aTHX != root) - return; -#endif + MUTEX_INIT(&su_uid_seq_counter_mutex); - SU_LOCK(&su_uid_seq_counter_mutex); - PerlMemShared_free(su_uid_seq_counter.seqs); + XSH_LOCK(&su_uid_seq_counter_mutex); + su_uid_seq_counter.seqs = NULL; su_uid_seq_counter.size = 0; - SU_UNLOCK(&su_uid_seq_counter_mutex); + XSH_UNLOCK(&su_uid_seq_counter_mutex); - MUTEX_DESTROY(&su_uid_seq_counter_mutex); - - su_initialized = 0; + stash = gv_stashpv(XSH_PACKAGE, 1); + newCONSTSUB(stash, "TOP", newSViv(0)); + newCONSTSUB(stash, "SU_THREADSAFE", newSVuv(XSH_THREADSAFE)); return; } -XS(XS_Scope__Upper_unwind); -XS(XS_Scope__Upper_yield); -XS(XS_Scope__Upper_leave); +static void xsh_user_local_setup(pTHX_ xsh_user_cxt_t *cxt) { + cxt->stack_placeholder = NULL; -#if SU_HAS_PERL(5, 9, 0) -# define SU_XS_FILE_TYPE const char -#else -# define SU_XS_FILE_TYPE char -#endif + /* NewOp() calls calloc() which just zeroes the memory with memset(). */ + Zero(&(cxt->unwind_storage.return_op), 1, LISTOP); + cxt->unwind_storage.return_op.op_type = OP_RETURN; + cxt->unwind_storage.return_op.op_ppaddr = PL_ppaddr[OP_RETURN]; -static void su_global_setup(pTHX_ SU_XS_FILE_TYPE *file) { -#define su_global_setup(F) su_global_setup(aTHX_ (F)) - HV *stash; + Zero(&(cxt->unwind_storage.proxy_op), 1, OP); + cxt->unwind_storage.proxy_op.op_type = OP_STUB; + cxt->unwind_storage.proxy_op.op_ppaddr = NULL; - if (su_initialized) - return; + Zero(&(cxt->yield_storage.leave_op), 1, UNOP); + cxt->yield_storage.leave_op.op_type = OP_STUB; + cxt->yield_storage.leave_op.op_ppaddr = NULL; - MUTEX_INIT(&su_uid_seq_counter_mutex); + Zero(&(cxt->yield_storage.proxy_op), 1, OP); + cxt->yield_storage.proxy_op.op_type = OP_STUB; + cxt->yield_storage.proxy_op.op_ppaddr = NULL; - SU_LOCK(&su_uid_seq_counter_mutex); - su_uid_seq_counter.seqs = NULL; - su_uid_seq_counter.size = 0; - SU_UNLOCK(&su_uid_seq_counter_mutex); + cxt->uplevel_storage.top = NULL; + cxt->uplevel_storage.root = NULL; + cxt->uplevel_storage.count = 0; - stash = gv_stashpv(__PACKAGE__, 1); - newCONSTSUB(stash, "TOP", newSViv(0)); - newCONSTSUB(stash, "SU_THREADSAFE", newSVuv(SU_THREADSAFE)); - - newXSproto("Scope::Upper::unwind", XS_Scope__Upper_unwind, file, NULL); - newXSproto("Scope::Upper::yield", XS_Scope__Upper_yield, file, NULL); - newXSproto("Scope::Upper::leave", XS_Scope__Upper_leave, file, NULL); - -#if SU_MULTIPLICITY - call_atexit(su_global_teardown, aTHX); -#else - call_atexit(su_global_teardown, NULL); -#endif - - su_initialized = 1; + cxt->uid_storage.map = NULL; + cxt->uid_storage.used = 0; + cxt->uid_storage.alloc = 0; return; } -/* --- Interpreter setup/teardown ------------------------------------------ */ - -static void su_local_teardown(pTHX_ void *param) { +static void xsh_user_local_teardown(pTHX_ xsh_user_cxt_t *cxt) { su_uplevel_ud *cur; - su_uid **map; - dMY_CXT; - map = MY_CXT.uid_storage.map; - if (map) { - STRLEN i; - for (i = 0; i < MY_CXT.uid_storage.used; ++i) - Safefree(map[i]); - Safefree(map); - } + Safefree(cxt->uid_storage.map); - cur = MY_CXT.uplevel_storage.root; + cur = cxt->uplevel_storage.root; if (cur) { su_uplevel_ud *prev; do { @@ -2364,38 +2321,13 @@ static void su_local_teardown(pTHX_ void *param) { return; } -static void su_local_setup(pTHX) { -#define su_local_setup() su_local_setup(aTHX) - MY_CXT_INIT; - - MY_CXT.stack_placeholder = NULL; - - /* NewOp() calls calloc() which just zeroes the memory with memset(). */ - Zero(&(MY_CXT.unwind_storage.return_op), 1, LISTOP); - MY_CXT.unwind_storage.return_op.op_type = OP_RETURN; - MY_CXT.unwind_storage.return_op.op_ppaddr = PL_ppaddr[OP_RETURN]; - - Zero(&(MY_CXT.unwind_storage.proxy_op), 1, OP); - MY_CXT.unwind_storage.proxy_op.op_type = OP_STUB; - MY_CXT.unwind_storage.proxy_op.op_ppaddr = NULL; - - Zero(&(MY_CXT.yield_storage.leave_op), 1, UNOP); - MY_CXT.yield_storage.leave_op.op_type = OP_STUB; - MY_CXT.yield_storage.leave_op.op_ppaddr = NULL; - - Zero(&(MY_CXT.yield_storage.proxy_op), 1, OP); - MY_CXT.yield_storage.proxy_op.op_type = OP_STUB; - MY_CXT.yield_storage.proxy_op.op_ppaddr = NULL; - - MY_CXT.uplevel_storage.top = NULL; - MY_CXT.uplevel_storage.root = NULL; - MY_CXT.uplevel_storage.count = 0; - - MY_CXT.uid_storage.map = NULL; - MY_CXT.uid_storage.used = 0; - MY_CXT.uid_storage.alloc = 0; +static void xsh_user_global_teardown(pTHX) { + XSH_LOCK(&su_uid_seq_counter_mutex); + PerlMemShared_free(su_uid_seq_counter.seqs); + su_uid_seq_counter.size = 0; + XSH_UNLOCK(&su_uid_seq_counter_mutex); - call_atexit(su_local_teardown, NULL); + MUTEX_DESTROY(&su_uid_seq_counter_mutex); return; } @@ -2432,7 +2364,7 @@ default_cx: \ } \ } STMT_END -#if SU_HAS_PERL(5, 10, 0) +#if XSH_HAS_PERL(5, 10, 0) # define SU_INFO_COUNT 11 #else # define SU_INFO_COUNT 10 @@ -2444,7 +2376,7 @@ XS(XS_Scope__Upper_unwind) { #else dXSARGS; #endif - dMY_CXT; + dXSH_CXT; I32 cxix; PERL_UNUSED_VAR(cv); /* -W */ @@ -2459,12 +2391,12 @@ XS(XS_Scope__Upper_unwind) { continue; case CXt_EVAL: case CXt_FORMAT: - MY_CXT.unwind_storage.cxix = cxix; - MY_CXT.unwind_storage.items = items; - MY_CXT.unwind_storage.savesp = PL_stack_sp; + XSH_CXT.unwind_storage.cxix = cxix; + XSH_CXT.unwind_storage.items = items; + XSH_CXT.unwind_storage.savesp = PL_stack_sp; if (items > 0) { - MY_CXT.unwind_storage.items--; - MY_CXT.unwind_storage.savesp--; + XSH_CXT.unwind_storage.items--; + XSH_CXT.unwind_storage.savesp--; } /* pp_entersub will want to sanitize the stack after returning from there * Screw that, we're insane! @@ -2488,19 +2420,19 @@ XS(XS_Scope__Upper_yield) { #else dXSARGS; #endif - dMY_CXT; + dXSH_CXT; I32 cxix; PERL_UNUSED_VAR(cv); /* -W */ PERL_UNUSED_VAR(ax); /* -Wall */ SU_GET_CONTEXT(0, items - 1, su_context_here()); - MY_CXT.yield_storage.cxix = cxix; - MY_CXT.yield_storage.items = items; - MY_CXT.yield_storage.savesp = PL_stack_sp; + XSH_CXT.yield_storage.cxix = cxix; + XSH_CXT.yield_storage.items = items; + XSH_CXT.yield_storage.savesp = PL_stack_sp; if (items > 0) { - MY_CXT.yield_storage.items--; - MY_CXT.yield_storage.savesp--; + XSH_CXT.yield_storage.items--; + XSH_CXT.yield_storage.savesp--; } /* See XS_Scope__Upper_unwind */ if (GIMME_V == G_SCALAR) @@ -2517,14 +2449,14 @@ XS(XS_Scope__Upper_leave) { #else dXSARGS; #endif - dMY_CXT; + dXSH_CXT; PERL_UNUSED_VAR(cv); /* -W */ PERL_UNUSED_VAR(ax); /* -Wall */ - MY_CXT.yield_storage.cxix = su_context_here(); - MY_CXT.yield_storage.items = items; - MY_CXT.yield_storage.savesp = PL_stack_sp; + XSH_CXT.yield_storage.cxix = su_context_here(); + XSH_CXT.yield_storage.items = items; + XSH_CXT.yield_storage.savesp = PL_stack_sp; /* See XS_Scope__Upper_unwind */ if (GIMME_V == G_SCALAR) PL_stack_sp = PL_stack_base + PL_markstack_ptr[1] + 1; @@ -2538,35 +2470,22 @@ PROTOTYPES: ENABLE BOOT: { - su_global_setup(file); - su_local_setup(); + xsh_setup(); + newXSproto("Scope::Upper::unwind", XS_Scope__Upper_unwind, file, NULL); + newXSproto("Scope::Upper::yield", XS_Scope__Upper_yield, file, NULL); + newXSproto("Scope::Upper::leave", XS_Scope__Upper_leave, file, NULL); } -#if SU_THREADSAFE +#if XSH_THREADSAFE void CLONE(...) PROTOTYPE: DISABLE -PREINIT: - su_uid_storage new_cxt; PPCODE: - { - dMY_CXT; - new_cxt.map = NULL; - new_cxt.used = 0; - new_cxt.alloc = 0; - su_uid_storage_dup(&new_cxt, &MY_CXT.uid_storage, MY_CXT.uid_storage.used); - } - { - MY_CXT_CLONE; - MY_CXT.uplevel_storage.top = NULL; - MY_CXT.uplevel_storage.root = NULL; - MY_CXT.uplevel_storage.count = 0; - MY_CXT.uid_storage = new_cxt; - } + xsh_clone(); XSRETURN(0); -#endif /* SU_THREADSAFE */ +#endif /* XSH_THREADSAFE */ void HERE() @@ -2617,6 +2536,7 @@ PPCODE: XSRETURN(1); } } + warn(su_no_such_target, "subroutine"); XSRETURN_UNDEF; void @@ -2637,6 +2557,7 @@ PPCODE: XSRETURN(1); } } + warn(su_no_such_target, "eval"); XSRETURN_UNDEF; void @@ -2789,7 +2710,7 @@ PPCODE: case CXt_EVAL: if (CxOLD_OP_TYPE(cx) == OP_ENTEREVAL) { /* eval STRING */ -#if SU_HAS_PERL(5, 17, 4) +#if XSH_HAS_PERL(5, 17, 4) PUSHs(newSVpvn_flags(SvPVX(cx->blk_eval.cur_text), SvCUR(cx->blk_eval.cur_text)-2, SvUTF8(cx->blk_eval.cur_text)|SVs_TEMP)); @@ -2816,7 +2737,7 @@ PPCODE: /* warnings (9) */ { SV *mask = NULL; -#if SU_HAS_PERL(5, 9, 4) +#if XSH_HAS_PERL(5, 9, 4) STRLEN *old_warnings = cop->cop_warnings; #else SV *old_warnings = cop->cop_warnings; @@ -2825,20 +2746,20 @@ PPCODE: if (PL_dowarn & G_WARN_ON) goto context_info_warnings_on; else -#if SU_HAS_PERL(5, 17, 4) +#if XSH_HAS_PERL(5, 17, 4) mask = &PL_sv_undef; #else goto context_info_warnings_off; #endif } else if (old_warnings == pWARN_NONE) { -#if !SU_HAS_PERL(5, 17, 4) +#if !XSH_HAS_PERL(5, 17, 4) context_info_warnings_off: #endif mask = su_newmortal_pvn(WARN_NONEstring, WARNsize); } else if (old_warnings == pWARN_ALL) { HV *bits; context_info_warnings_on: -#if SU_HAS_PERL(5, 8, 7) +#if XSH_HAS_PERL(5, 8, 7) bits = get_hv("warnings::Bits", 0); if (bits) { SV **bits_all = hv_fetchs(bits, "all", FALSE); @@ -2849,7 +2770,7 @@ context_info_warnings_on: if (!mask) mask = su_newmortal_pvn(WARN_ALLstring, WARNsize); } else { -#if SU_HAS_PERL(5, 9, 4) +#if XSH_HAS_PERL(5, 9, 4) mask = su_newmortal_pvn((char *) (old_warnings + 1), old_warnings[0]); #else mask = sv_mortalcopy(old_warnings); @@ -2857,7 +2778,7 @@ context_info_warnings_on: } PUSHs(mask); } -#if SU_HAS_PERL(5, 10, 0) +#if XSH_HAS_PERL(5, 10, 0) /* hints hash (10) */ { COPHH *hints_hash = CopHINTHASH_get(cop);