kernel/eka/klib/x86/cumem.cia
changeset 0 a41df078684a
equal deleted inserted replaced
-1:000000000000 0:a41df078684a
       
     1 // Copyright (c) 2007-2009 Nokia Corporation and/or its subsidiary(-ies).
       
     2 // All rights reserved.
       
     3 // This component and the accompanying materials are made available
       
     4 // under the terms of the License "Eclipse Public License v1.0"
       
     5 // which accompanies this distribution, and is available
       
     6 // at the URL "http://www.eclipse.org/legal/epl-v10.html".
       
     7 //
       
     8 // Initial Contributors:
       
     9 // Nokia Corporation - initial contribution.
       
    10 //
       
    11 // Contributors:
       
    12 //
       
    13 // Description:
       
    14 // e32\klib\x86\cumem.cia
       
    15 // 
       
    16 //
       
    17 
       
    18 #include <x86.h>
       
    19 
       
    20 extern "C" {
       
    21 
       
    22 __NAKED__ void CopyInterSeg()
       
    23 //
       
    24 // Copy ECX bytes from DS:ESI to ES:EDI
       
    25 // Modifies EAX, EBX, ECX, EDX, ESI, EDI
       
    26 //
       
    27 	{
       
    28 	asm("pushfd");
       
    29 	asm("cld");				// assume forward copy initially
       
    30 	asm("test ecx,ecx");	//
       
    31 	asm("jz short memcopy0");// if length=0, nothing to do
       
    32 	asm("xor edx,edx");		//
       
    33 	asm("cmp edi,esi");		// compare source and dest addresses
       
    34 	asm("jc short memcopy1");// if dest<source, must go forwards
       
    35 	asm("std");				// else go backwards
       
    36 	asm("add esi,ecx");		// and start at end of block
       
    37 	asm("add edi,ecx");		//
       
    38 	asm("inc edx");			// edx=1 if backwards, 0 if forwards
       
    39 	asm("memcopy1:");
       
    40 	asm("cmp ecx,16");		// if length<16 don't bother with alignment check
       
    41 	asm("jc short memcopy2");//
       
    42 	asm("mov ebx,edi");		// ebx = destination address
       
    43 	asm("and ebx,3");		// ebx bottom 2 bits = alignment of destination wrt hardware bus
       
    44 	asm("jz short memcopy3");// if aligned, proceed with block move
       
    45 	asm("or edx,edx");		// check direction of move
       
    46 	asm("jnz short memcopy4");// if backwards, ebx = number of byte moves to align destination
       
    47 	asm("neg ebx");			// else number of byte moves = 4-ebx
       
    48 	asm("add ebx,4");		//
       
    49 	asm("memcopy4:");
       
    50 	asm("sub ecx,ebx");		// subtract number of bytes from length
       
    51 	asm("xchg ecx,ebx");	// temporarily put length in ebx
       
    52 	asm("sub edi,edx");		// adjust if backwards move
       
    53 	asm("sub esi,edx");		//
       
    54 	asm("rep movsb");		// move bytes to align destination
       
    55 	asm("add edi,edx");		// adjust if backwards move
       
    56 	asm("add esi,edx");		//
       
    57 	asm("mov ecx,ebx");		// length back into ecx
       
    58 	asm("memcopy3:");
       
    59 	asm("mov ebx,ecx");		// save length in ebx
       
    60 	asm("shl edx,2");		// adjustment 4 for backwards move
       
    61 	asm("shr ecx,2");		// number of dwords to move into ecx
       
    62 	asm("sub edi,edx");		// adjust if backwards move
       
    63 	asm("sub esi,edx");		//
       
    64 	asm("rep movsd");		// perform DWORD block move
       
    65 	asm("add edi,edx");		// adjust if backwards move
       
    66 	asm("add esi,edx");		//
       
    67 	asm("mov ecx,ebx");		// length back into ecx
       
    68 	asm("and ecx,3");		// number of remaining bytes to move
       
    69 	asm("jz short memcopy0");// if zero, we are finished
       
    70 	asm("shr edx,2");		// adjustment 1 for backwards move
       
    71 	asm("memcopy2:");		// *** come here for small move
       
    72 	asm("sub edi,edx");		// adjust if backwards move
       
    73 	asm("sub esi,edx");		//
       
    74 	asm("rep movsb");		// move remaining bytes
       
    75 	asm("memcopy0:");
       
    76 	asm("popfd");
       
    77 	asm("ret");				// finished - return value in EAX
       
    78 	}
       
    79 
       
    80 __NAKED__ void CopyInterSeg32()
       
    81 //
       
    82 // Copy ECX bytes from DS:ESI to ES:EDI
       
    83 // ECX, ESI and EDI are all multiples of 4
       
    84 // Modifies EAX, EBX, ECX, EDX, ESI, EDI
       
    85 //
       
    86 	{
       
    87 	asm("pushfd");
       
    88 	asm("cld");				//
       
    89 	asm("test ecx,ecx");	//
       
    90 	asm("jz short memmove0");// if length=0, nothing to do
       
    91 	asm("cmp edi,esi");		// compare source and dest addresses
       
    92 	asm("jc short memmove1");// if dest<source, must go forwards
       
    93 	asm("std");				// else go backwards
       
    94 	asm("lea esi,[esi+ecx-4]");	// and start at end of block - 4
       
    95 	asm("lea edi,[edi+ecx-4]");	//
       
    96 	asm("memmove1:");
       
    97 	asm("shr ecx,2");		// ecx now contains number of dwords to move
       
    98 	asm("rep movsd");		// do dword block move
       
    99 	asm("memmove0:");
       
   100 	asm("popfd");
       
   101 	asm("ret");				// finished - return value in EAX
       
   102 	}
       
   103 
       
   104 __NAKED__ void FillInterSeg()
       
   105 //
       
   106 // Fill ECX bytes at ES:EDI with AL
       
   107 // Modifies EAX, ECX, EDX, EDI
       
   108 //
       
   109 	{
       
   110 	asm("pushfd");
       
   111 	asm("cld");				// go forwards through array
       
   112 	asm("test ecx,ecx");	//
       
   113 	asm("jz short memfill0");// if length zero, nothing to do
       
   114 	asm("cmp ecx,8");			// if array very small, just do byte fills
       
   115 	asm("jb short memfill1");
       
   116 
       
   117 	asm("mov ah,al");		// repeat al in all bytes of eax
       
   118 	asm("movzx edx,ax");	//
       
   119 	asm("shl eax,16");		//
       
   120 	asm("or eax,edx");		//
       
   121 	asm("mov edx,ecx");		// length into edx
       
   122 	// ecx = number of byte fills to align = 4-(edi mod 4)
       
   123 	asm("mov ecx,4");		
       
   124 	asm("sub ecx,edi");		//
       
   125 	asm("and ecx,3");		//
       
   126 	asm("jz short memfill2");// if already aligned, proceed to dword fill
       
   127 	asm("sub edx,ecx");		// subtract alignment bytes from length
       
   128 	asm("rep stosb");		// do byte fills to align
       
   129 	asm("memfill2:");
       
   130 	asm("mov ecx,edx");		// length remaining into ecx
       
   131 	asm("shr ecx,2");		// number of dwords to fill into ecx
       
   132 	asm("rep stosd");		// perform dword fill
       
   133 	asm("mov ecx,edx");		// calculate number of leftover bytes
       
   134 	asm("and ecx,3");		// in ecx
       
   135 	asm("jz short memfill0");// if none left, exit
       
   136 	asm("memfill1:");
       
   137 	asm("rep stosb");		// do byte fills to make up correct length
       
   138 	asm("memfill0:");
       
   139 	asm("popfd");
       
   140 	asm("ret");
       
   141 	}
       
   142 
       
   143 
       
   144 /**	Reads the current thread's memory space with appropriate permissions.
       
   145 
       
   146 Performs a memcpy(aKernAddr, aAddr, aLength).
       
   147 The reads are performed	using requestor privilege level from GS, ie equal
       
   148 to the privilege level of the caller of the Exec:: function.
       
   149 Note that source and destination areas may not overlap.
       
   150 
       
   151 @param	aKernAddr	Destination address in kernel memory.
       
   152 @param	aAddr		Source address in kernel or user memory.
       
   153 @param	aLength		Number of bytes to copy.
       
   154 
       
   155 @pre    Call in a thread context.
       
   156 @pre    Kernel must be unlocked.
       
   157 @pre	Must be called under an XTRAP harness, or calling thread must not be
       
   158 in a critical section.
       
   159 */
       
   160 EXPORT_C __NAKED__ void kumemget(TAny* /*aKernAddr*/, const TAny* /*aAddr*/, TInt /*aLength*/)
       
   161 	{
       
   162 	asm("push edi");
       
   163 	asm("push esi");
       
   164 	asm("push ebx");
       
   165 	asm("push ds");
       
   166 	asm("mov edi, [esp+20]");
       
   167 	asm("mov esi, [esp+24]");
       
   168 	asm("mov ecx, [esp+28]");
       
   169 	asm("mov ax, gs");
       
   170 	asm("mov ds, ax");
       
   171 	asm("call %a0": : "i"(&CopyInterSeg));
       
   172 	asm("pop ds");
       
   173 	asm("pop ebx");
       
   174 	asm("pop esi");
       
   175 	asm("pop edi");
       
   176 	asm("ret");
       
   177 	}
       
   178 
       
   179 
       
   180 /**	Reads the current thread's memory space with user permissions.
       
   181 
       
   182 Performs a memcpy(aKernAddr, aUserAddr, aLength).
       
   183 The reads are performed with ring 3 RPL.
       
   184 Note that source and destination areas may not overlap.
       
   185 
       
   186 @param	aKernAddr	Destination address in kernel memory.
       
   187 @param	aUserAddr	Source address in user memory.
       
   188 @param	aLength		Number of bytes to copy.
       
   189 
       
   190 @pre    Call in a thread context.
       
   191 @pre    Kernel must be unlocked.
       
   192 @pre	Must be called under an XTRAP harness, or calling thread must not be
       
   193 in a critical section.
       
   194 */
       
   195 EXPORT_C __NAKED__ void umemget(TAny* /*aKernAddr*/, const TAny* /*aUserAddr*/, TInt /*aLength*/)
       
   196 	{
       
   197 	asm("push edi");
       
   198 	asm("push esi");
       
   199 	asm("push ebx");
       
   200 	asm("push ds");
       
   201 	asm("mov edi, [esp+20]");
       
   202 	asm("mov esi, [esp+24]");
       
   203 	asm("mov ecx, [esp+28]");
       
   204 	asm("mov eax, %0": : "i"(RING3_DS));
       
   205 	asm("mov ds, ax");
       
   206 	asm("call %a0": : "i"(&CopyInterSeg));
       
   207 	asm("pop ds");
       
   208 	asm("pop ebx");
       
   209 	asm("pop esi");
       
   210 	asm("pop edi");
       
   211 	asm("ret");
       
   212 	}
       
   213 
       
   214 
       
   215 /**	Does a word-aligned read of the current thread's memory space with appropriate permissions.
       
   216 
       
   217 Performs a memcpy(aKernAddr, aAddr, aLength).
       
   218 The reads are performed	using requestor privilege level from GS, ie equal
       
   219 to the privilege level of the caller of the Exec:: function.
       
   220 Note that source and destination areas may not overlap.
       
   221 
       
   222 @param	aKernAddr	Destination address in kernel memory, must be 4-byte aligned.
       
   223 @param	aAddr		Source address in kernel or user memory, must be 4-byte aligned.
       
   224 @param	aLength		Number of bytes to copy, must be a multiple of 4.
       
   225 
       
   226 @pre    Call in a thread context.
       
   227 @pre    Kernel must be unlocked.
       
   228 @pre	Must be called under an XTRAP harness, or calling thread must not be
       
   229 in a critical section.
       
   230 */
       
   231 EXPORT_C __NAKED__ void kumemget32(TAny* /*aKernAddr*/, const TAny* /*aAddr*/, TInt /*aLength*/)
       
   232 	{
       
   233 	asm("push edi");
       
   234 	asm("push esi");
       
   235 	asm("push ebx");
       
   236 	asm("push ds");
       
   237 	asm("mov edi, [esp+20]");
       
   238 	asm("mov esi, [esp+24]");
       
   239 	asm("mov ecx, [esp+28]");
       
   240 	asm("mov ax, gs");
       
   241 	asm("mov ds, ax");
       
   242 	asm("call %a0": : "i"(&CopyInterSeg32));
       
   243 	asm("pop ds");
       
   244 	asm("pop ebx");
       
   245 	asm("pop esi");
       
   246 	asm("pop edi");
       
   247 	asm("ret");
       
   248 	}
       
   249 
       
   250 
       
   251 /**	Does a word-aligned read of the current thread's memory space with user permissions.
       
   252 
       
   253 Performs a memcpy(aKernAddr, aUserAddr, aLength).
       
   254 The reads are performed with ring 3 RPL.
       
   255 Note that source and destination areas may not overlap.
       
   256 
       
   257 @param	aKernAddr	Destination address in kernel memory, must be 4-byte aligned.
       
   258 @param	aUserAddr	Source address in user memory, must be 4-byte aligned.
       
   259 @param	aLength		Number of bytes to copy, must be a multiple of 4.
       
   260 
       
   261 @pre    Call in a thread context.
       
   262 @pre    Kernel must be unlocked.
       
   263 @pre	Must be called under an XTRAP harness, or calling thread must not be
       
   264 in a critical section.
       
   265 */
       
   266 EXPORT_C __NAKED__ void umemget32(TAny* /*aKernAddr*/, const TAny* /*aUserAddr*/, TInt /*aLength*/)
       
   267 	{
       
   268 	asm("push edi");
       
   269 	asm("push esi");
       
   270 	asm("push ebx");
       
   271 	asm("push ds");
       
   272 	asm("mov edi, [esp+20]");
       
   273 	asm("mov esi, [esp+24]");
       
   274 	asm("mov ecx, [esp+28]");
       
   275 	asm("mov eax, %0": : "i"(RING3_DS));
       
   276 	asm("mov ds, ax");
       
   277 	asm("call %a0": : "i"(&CopyInterSeg32));
       
   278 	asm("pop ds");
       
   279 	asm("pop ebx");
       
   280 	asm("pop esi");
       
   281 	asm("pop edi");
       
   282 	asm("ret");
       
   283 	}
       
   284 
       
   285 
       
   286 /**	Writes to the current thread's memory space with appropriate permissions.
       
   287 
       
   288 Performs a memcpy(aAddr, aKernAddr, aLength).
       
   289 The writes are performed using requestor privilege level from GS, ie equal
       
   290 to the privilege level of the caller of the Exec:: function.
       
   291 Note that source and destination areas may not overlap.
       
   292 
       
   293 @param	aAddr		Destination address in kernel or user memory.
       
   294 @param	aKernAddr	Source address in kernel memory.
       
   295 @param	aLength		Number of bytes to copy.
       
   296 
       
   297 @pre    Call in a thread context.
       
   298 @pre    Kernel must be unlocked.
       
   299 @pre	Must be called under an XTRAP harness, or calling thread must not be
       
   300 in a critical section.
       
   301 */
       
   302 EXPORT_C __NAKED__ void kumemput(TAny* /*aAddr*/, const TAny* /*aKernAddr*/, TInt /*aLength*/)
       
   303 	{
       
   304 	asm("push edi");
       
   305 	asm("push esi");
       
   306 	asm("push ebx");
       
   307 	asm("push es");
       
   308 	asm("mov edi, [esp+20]");
       
   309 	asm("mov esi, [esp+24]");
       
   310 	asm("mov ecx, [esp+28]");
       
   311 	asm("mov ax, gs");
       
   312 	asm("mov es, ax");
       
   313 	asm("call %a0": : "i"(&CopyInterSeg));
       
   314 	asm("pop es");
       
   315 	asm("pop ebx");
       
   316 	asm("pop esi");
       
   317 	asm("pop edi");
       
   318 	asm("ret");
       
   319 	}
       
   320 
       
   321 
       
   322 /**	Writes to the current thread's memory space with user permissions.
       
   323 
       
   324 Performs a memcpy(aAddr, aKernAddr, aLength).
       
   325 The writes are performed with ring 3 RPL.
       
   326 Note that source and destination areas may not overlap.
       
   327 
       
   328 @param	aUserAddr	Destination address in user memory.
       
   329 @param	aKernAddr	Source address in kernel memory.
       
   330 @param	aLength		Number of bytes to copy.
       
   331 
       
   332 @pre    Call in a thread context.
       
   333 @pre    Kernel must be unlocked.
       
   334 @pre	Must be called under an XTRAP harness, or calling thread must not be
       
   335 in a critical section.
       
   336 */
       
   337 EXPORT_C __NAKED__ void umemput(TAny* /*aUserAddr*/, const TAny* /*aKernAddr*/, TInt /*aLength*/)
       
   338 	{
       
   339 	asm("push edi");
       
   340 	asm("push esi");
       
   341 	asm("push ebx");
       
   342 	asm("push es");
       
   343 	asm("mov edi, [esp+20]");
       
   344 	asm("mov esi, [esp+24]");
       
   345 	asm("mov ecx, [esp+28]");
       
   346 	asm("mov eax, %0": : "i"(RING3_DS));
       
   347 	asm("mov es, ax");
       
   348 	asm("call %a0": : "i"(&CopyInterSeg));
       
   349 	asm("pop es");
       
   350 	asm("pop ebx");
       
   351 	asm("pop esi");
       
   352 	asm("pop edi");
       
   353 	asm("ret");
       
   354 	}
       
   355 
       
   356 
       
   357 /**	Does a word-aligned write to the current thread's memory space with appropriate permissions.
       
   358 
       
   359 Performs a memcpy(aAddr, aKernAddr, aLength).
       
   360 The writes are performed using requestor privilege level from GS, ie equal
       
   361 to the privilege level of the caller of the Exec:: function.
       
   362 Note that source and destination areas may not overlap.
       
   363 
       
   364 @param	aAddr		Destination address in kernel or user memory, must be 4-byte aligned.
       
   365 @param	aKernAddr	Source address in kernel memory, must be 4-byte aligned.
       
   366 @param	aLength		Number of bytes to copy, must be a multiple of 4.
       
   367 
       
   368 @pre    Call in a thread context.
       
   369 @pre    Kernel must be unlocked.
       
   370 @pre	Must be called under an XTRAP harness, or calling thread must not be
       
   371 in a critical section.
       
   372 */
       
   373 EXPORT_C __NAKED__ void kumemput32(TAny* /*aAddr*/, const TAny* /*aKernAddr*/, TInt /*aLength*/)
       
   374 	{
       
   375 	asm("push edi");
       
   376 	asm("push esi");
       
   377 	asm("push ebx");
       
   378 	asm("push es");
       
   379 	asm("mov edi, [esp+20]");
       
   380 	asm("mov esi, [esp+24]");
       
   381 	asm("mov ecx, [esp+28]");
       
   382 	asm("mov ax, gs");
       
   383 	asm("mov es, ax");
       
   384 	asm("call %a0": : "i"(&CopyInterSeg32));
       
   385 	asm("pop es");
       
   386 	asm("pop ebx");
       
   387 	asm("pop esi");
       
   388 	asm("pop edi");
       
   389 	asm("ret");
       
   390 	}
       
   391 
       
   392 
       
   393 /**	Does a word-aligned write to the current thread's memory space with user permissions.
       
   394 
       
   395 Performs a memcpy(aAddr, aKernAddr, aLength).
       
   396 The writes are performed with ring 3 RPL.
       
   397 Note that source and destination areas may not overlap.
       
   398 
       
   399 @param	aUserAddr	Destination address in user memory, must be 4-byte aligned.
       
   400 @param	aKernAddr	Source address in kernel memory, must be 4-byte aligned.
       
   401 @param	aLength		Number of bytes to copy, must be a multiple of 4.
       
   402 
       
   403 @pre    Call in a thread context.
       
   404 @pre    Kernel must be unlocked.
       
   405 @pre	Must be called under an XTRAP harness, or calling thread must not be
       
   406 in a critical section.
       
   407 */
       
   408 EXPORT_C __NAKED__ void umemput32(TAny* /*aUserAddr*/, const TAny* /*aKernAddr*/, TInt /*aLength*/)
       
   409 	{
       
   410 	asm("push edi");
       
   411 	asm("push esi");
       
   412 	asm("push ebx");
       
   413 	asm("push es");
       
   414 	asm("mov edi, [esp+20]");
       
   415 	asm("mov esi, [esp+24]");
       
   416 	asm("mov ecx, [esp+28]");
       
   417 	asm("mov eax, %0": : "i"(RING3_DS));
       
   418 	asm("mov es, ax");
       
   419 	asm("call %a0": : "i"(&CopyInterSeg32));
       
   420 	asm("pop es");
       
   421 	asm("pop ebx");
       
   422 	asm("pop esi");
       
   423 	asm("pop edi");
       
   424 	asm("ret");
       
   425 	}
       
   426 
       
   427 
       
   428 /**	Fills the current thread's memory space with appropriate permissions.
       
   429 
       
   430 Performs a memset(aAddr, aValue, aLength).
       
   431 The writes are performed using requestor privilege level from GS, ie equal
       
   432 to the privilege level of the caller of the Exec:: function.
       
   433 
       
   434 @param	aAddr		Destination address in kernel or user memory.
       
   435 @param	aValue		Value to write to each byte.
       
   436 @param	aLength		Number of bytes to fill.
       
   437 
       
   438 @pre    Call in a thread context.
       
   439 @pre    Kernel must be unlocked.
       
   440 @pre	Must be called under an XTRAP harness, or calling thread must not be
       
   441 in a critical section.
       
   442 */
       
   443 EXPORT_C __NAKED__ void kumemset(TAny* /*aAddr*/, const TUint8 /*aValue*/, TInt /*aLength*/)
       
   444 	{
       
   445 	asm("push edi");
       
   446 	asm("push es");
       
   447 	asm("mov edi, [esp+12]");
       
   448 	asm("mov eax, [esp+16]");
       
   449 	asm("mov ecx, [esp+20]");
       
   450 	asm("mov dx, gs");
       
   451 	asm("mov es, dx");
       
   452 	asm("call %a0": :"i"(&FillInterSeg));
       
   453 	asm("pop es");
       
   454 	asm("pop edi");
       
   455 	asm("ret");
       
   456 	}
       
   457 
       
   458 
       
   459 /**	Fills the current thread's memory space with user permissions.
       
   460 
       
   461 Performs a memset(aUserAddr, aValue, aLength).
       
   462 The writes are performed with ring 3 RPL.
       
   463 
       
   464 @param	aUserAddr	Destination address in user memory.
       
   465 @param	aValue		Value to write to each byte.
       
   466 @param	aLength		Number of bytes to fill.
       
   467 
       
   468 @pre    Call in a thread context.
       
   469 @pre    Kernel must be unlocked.
       
   470 @pre	Must be called under an XTRAP harness, or calling thread must not be
       
   471 in a critical section.
       
   472 */
       
   473 EXPORT_C __NAKED__ void umemset(TAny* /*aUserAddr*/, const TUint8 /*aValue*/, TInt /*aLength*/)
       
   474 	{
       
   475 	asm("push edi");
       
   476 	asm("push es");
       
   477 	asm("mov edi, [esp+12]");
       
   478 	asm("mov eax, [esp+16]");
       
   479 	asm("mov ecx, [esp+20]");
       
   480 	asm("mov edx, %0": : "i"(RING3_DS));
       
   481 	asm("mov es, dx");
       
   482 	asm("call %a0": :"i"(&FillInterSeg));
       
   483 	asm("pop es");
       
   484 	asm("pop edi");
       
   485 	asm("ret");
       
   486 	}
       
   487 
       
   488 __NAKED__ void uumemcpy32(TAny* /*aUserDst*/, const TAny* /*aUserSrc*/, TInt /*aLength*/)
       
   489 	{
       
   490 	asm("push edi");
       
   491 	asm("push esi");
       
   492 	asm("push ebx");
       
   493 	asm("push ds");
       
   494 	asm("push es");
       
   495 	asm("mov edi, [esp+24]");
       
   496 	asm("mov esi, [esp+28]");
       
   497 	asm("mov ecx, [esp+32]");
       
   498 	asm("mov ax, gs");
       
   499 	asm("mov ds, ax");
       
   500 	asm("mov es, ax");
       
   501 	asm("call %a0": : "i"(&CopyInterSeg32));
       
   502 	asm("pop es");
       
   503 	asm("pop ds");
       
   504 	asm("pop ebx");
       
   505 	asm("pop esi");
       
   506 	asm("pop edi");
       
   507 	asm("ret");
       
   508 	}
       
   509 
       
   510 __NAKED__ void uumemcpy(TAny* /*aUserDst*/, const TAny* /*aUserSrc*/, TInt /*aLength*/)
       
   511 	{
       
   512 	asm("push edi");
       
   513 	asm("push esi");
       
   514 	asm("push ebx");
       
   515 	asm("push ds");
       
   516 	asm("push es");
       
   517 	asm("mov edi, [esp+24]");
       
   518 	asm("mov esi, [esp+28]");
       
   519 	asm("mov ecx, [esp+32]");
       
   520 	asm("mov ax, gs");
       
   521 	asm("mov ds, ax");
       
   522 	asm("mov es, ax");
       
   523 	asm("call %a0": : "i"(&CopyInterSeg));
       
   524 	asm("pop es");
       
   525 	asm("pop ds");
       
   526 	asm("pop ebx");
       
   527 	asm("pop esi");
       
   528 	asm("pop edi");
       
   529 	asm("ret");
       
   530 	}
       
   531 
       
   532 }