|
1 // Copyright (c) 2009 Nokia Corporation and/or its subsidiary(-ies). |
|
2 // All rights reserved. |
|
3 // This component and the accompanying materials are made available |
|
4 // under the terms of the License "Eclipse Public License v1.0" |
|
5 // which accompanies this distribution, and is available |
|
6 // at the URL "http://www.eclipse.org/legal/epl-v10.html". |
|
7 // |
|
8 // Initial Contributors: |
|
9 // Nokia Corporation - initial contribution. |
|
10 // |
|
11 // Contributors: |
|
12 // |
|
13 // Description: |
|
14 // e32/memmodel/emul/win32/mshbuf.cpp |
|
15 // Shareable Data Buffers |
|
16 |
|
17 #include "memmodel.h" |
|
18 #include <kernel/smap.h> |
|
19 |
|
20 _LIT(KLitDWin32ShPool,"DWin32ShPool"); |
|
21 _LIT(KLitDWin32AlignedShPool,"DWin32AlignedShPool"); |
|
22 _LIT(KLitDWin32NonAlignedShPool,"DWin32NonAlignedShPool"); |
|
23 |
|
24 |
|
25 DWin32ShBuf::DWin32ShBuf(DShPool* aPool, TLinAddr aRelAddr) : DShBuf(aPool, aRelAddr) |
|
26 { |
|
27 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32ShBuf::DWin32ShBuf()")); |
|
28 } |
|
29 |
|
30 DWin32ShBuf::~DWin32ShBuf() |
|
31 { |
|
32 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32ShBuf::~DWin32ShBuf()")); |
|
33 } |
|
34 |
|
35 TUint8* DWin32ShBuf::Base(DProcess* aProcess) |
|
36 { |
|
37 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32ShBuf::Base(0x%x)", aProcess)); |
|
38 |
|
39 TUint8* base = reinterpret_cast<DWin32ShPool*>(iPool)->Base(aProcess) + (TUint)iRelAddress; |
|
40 |
|
41 return base; |
|
42 } |
|
43 |
|
44 TUint8* DWin32ShBuf::Base() |
|
45 { |
|
46 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32ShBuf::Base()")); |
|
47 |
|
48 TUint8* base = reinterpret_cast<DWin32ShPool*>(iPool)->Base() + (TUint)iRelAddress; |
|
49 |
|
50 return base; |
|
51 } |
|
52 |
|
53 TInt DWin32ShBuf::Map(TUint /* aMapAttr */, DProcess* /* aProcess */, TLinAddr& aBase) |
|
54 { |
|
55 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32ShBuf::Map()")); |
|
56 |
|
57 TInt r = KErrNotSupported; |
|
58 |
|
59 if (iPool->iPoolFlags & EShPoolPageAlignedBuffer) |
|
60 { |
|
61 if(iMapped) |
|
62 { |
|
63 r = KErrAlreadyExists; |
|
64 } |
|
65 else |
|
66 { |
|
67 aBase = reinterpret_cast<TUint>(reinterpret_cast<DWin32ShPool*>(iPool)->Base() + (TUint)iRelAddress); |
|
68 iMapped = ETrue; |
|
69 r = KErrNone; |
|
70 } |
|
71 } |
|
72 |
|
73 return r; |
|
74 } |
|
75 |
|
76 TInt DWin32ShBuf::UnMap(DProcess* /* aProcess */) |
|
77 { |
|
78 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32ShBuf::UnMap()")); |
|
79 |
|
80 TInt r = KErrNotSupported; |
|
81 |
|
82 if (iPool->iPoolFlags & EShPoolPageAlignedBuffer) |
|
83 { |
|
84 if(iMapped) |
|
85 { |
|
86 iMapped = EFalse; |
|
87 r = KErrNone; |
|
88 } |
|
89 else |
|
90 { |
|
91 r = KErrNotFound; |
|
92 } |
|
93 } |
|
94 |
|
95 return r; |
|
96 } |
|
97 |
|
98 TInt DWin32ShBuf::AddToProcess(DProcess* aProcess, TUint /* aAttr */) |
|
99 { |
|
100 __KTRACE_OPT(KMMU, Kern::Printf("Adding DWin32ShBuf %O to process %O", this, aProcess)); |
|
101 TUint flags; |
|
102 TInt r = KErrNone; |
|
103 |
|
104 if (aProcess != K::TheKernelProcess) |
|
105 r = iPool->OpenClient(aProcess, flags); |
|
106 |
|
107 return r; |
|
108 } |
|
109 |
|
110 TInt DWin32ShBuf::Close(TAny* aPtr) |
|
111 { |
|
112 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32ShBuf::Close(0x%08x)", aPtr)); |
|
113 |
|
114 if (aPtr) |
|
115 { |
|
116 DProcess* pP = reinterpret_cast<DProcess*>(aPtr); |
|
117 |
|
118 if (pP != K::TheKernelProcess) |
|
119 iPool->CloseClient(pP); |
|
120 } |
|
121 |
|
122 return DShBuf::Close(aPtr); |
|
123 } |
|
124 |
|
125 DWin32ShPool::DWin32ShPool() |
|
126 : DShPool() |
|
127 { |
|
128 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32ShPool::DWin32ShPool")); |
|
129 } |
|
130 |
|
131 |
|
132 DWin32ShPool::~DWin32ShPool() |
|
133 { |
|
134 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32ShPool::~DWin32ShPool")); |
|
135 |
|
136 if (iWin32MemoryBase) |
|
137 { |
|
138 TUint64 maxSize = static_cast<TUint64>(iMaxBuffers) * static_cast<TUint64>(iBufGap); |
|
139 |
|
140 // We know that maxSize is less than KMaxTInt as we tested for this in DoCreate(). |
|
141 VirtualFree(LPVOID(iWin32MemoryBase), (SIZE_T)maxSize, MEM_DECOMMIT); |
|
142 VirtualFree(LPVOID(iWin32MemoryBase), 0, MEM_RELEASE); |
|
143 MM::Wait(); |
|
144 MM::FreeMemory += iWin32MemorySize; |
|
145 MM::Signal(); |
|
146 } |
|
147 |
|
148 delete iBufMap; |
|
149 } |
|
150 |
|
151 void DWin32ShPool::DestroyClientResources(DProcess* aProcess) |
|
152 { |
|
153 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32ShPool::DestroyClientResources")); |
|
154 |
|
155 TInt r = DestroyHandles(aProcess); |
|
156 __NK_ASSERT_DEBUG((r == KErrNone) || (r == KErrDied)); |
|
157 (void)r; // Silence warnings |
|
158 } |
|
159 |
|
160 TInt DWin32ShPool::DeleteInitialBuffers() |
|
161 { |
|
162 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32ShPool::DeleteInitialBuffers")); |
|
163 |
|
164 if (iInitialBuffersArray != NULL) |
|
165 { |
|
166 for (TUint i = 0; i < iInitialBuffers; i++) |
|
167 { |
|
168 iInitialBuffersArray[i].iObjLink.Deque(); // remove from free list |
|
169 iInitialBuffersArray[i].Dec(); |
|
170 iInitialBuffersArray[i].~DWin32ShBuf(); |
|
171 } |
|
172 |
|
173 Kern::Free(iInitialBuffersArray); |
|
174 iInitialBuffersArray = NULL; |
|
175 } |
|
176 |
|
177 return KErrNone; |
|
178 } |
|
179 |
|
180 TInt DWin32ShPool::DestroyHandles(DProcess* aProcess) |
|
181 { |
|
182 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32ShPool::DestroyHandles(0x%08x)", aProcess)); |
|
183 |
|
184 TInt r = KErrNone; |
|
185 Kern::MutexWait(*iProcessLock); |
|
186 DShPoolClient* client = reinterpret_cast<DShPoolClient*>(iClientMap->Remove(reinterpret_cast<TUint>(aProcess))); |
|
187 |
|
188 __NK_ASSERT_DEBUG(client); |
|
189 __NK_ASSERT_DEBUG(client->iAccessCount == 0); |
|
190 |
|
191 delete client; |
|
192 |
|
193 if (aProcess != K::TheKernelProcess) |
|
194 { |
|
195 // Remove reserved handles |
|
196 r = aProcess->iHandles.Reserve(-TInt(iTotalBuffers)); |
|
197 } |
|
198 |
|
199 Kern::MutexSignal(*iProcessLock); |
|
200 |
|
201 return r; |
|
202 } |
|
203 |
|
204 |
|
205 TInt DWin32ShPool::Close(TAny* aPtr) |
|
206 { |
|
207 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32ShPool::Close(0x%08x)", aPtr)); |
|
208 |
|
209 if (aPtr) // not NULL must be user side |
|
210 { |
|
211 DProcess* pP = reinterpret_cast<DProcess*>(aPtr); |
|
212 |
|
213 CloseClient(pP); |
|
214 } |
|
215 |
|
216 return DShPool::Close(aPtr); |
|
217 } |
|
218 |
|
219 |
|
220 TInt DWin32ShPool::CreateInitialBuffers() |
|
221 { |
|
222 __KTRACE_OPT(KMMU,Kern::Printf(">DWin32ShPool::CreateInitialBuffers")); |
|
223 |
|
224 iInitialBuffersArray = reinterpret_cast<DWin32ShBuf*>(Kern::Alloc(iInitialBuffers * sizeof(DWin32ShBuf))); |
|
225 |
|
226 if (iInitialBuffersArray == NULL) |
|
227 return KErrNoMemory; |
|
228 |
|
229 TLinAddr offset = 0; |
|
230 for (TUint i = 0; i < iInitialBuffers; i++) |
|
231 { |
|
232 DWin32ShBuf *buf = new (&iInitialBuffersArray[i]) DWin32ShBuf(this, offset); |
|
233 TInt r = buf->Construct(); |
|
234 |
|
235 if (r == KErrNone) |
|
236 { |
|
237 iFreeList.Add(&buf->iObjLink); |
|
238 } |
|
239 else |
|
240 { |
|
241 iInitialBuffers = i; |
|
242 return KErrNoMemory; |
|
243 } |
|
244 |
|
245 offset += iBufGap; |
|
246 } |
|
247 |
|
248 iFreeBuffers = iInitialBuffers; |
|
249 iTotalBuffers = iInitialBuffers; |
|
250 |
|
251 iBufMap->Alloc(0, iInitialBuffers); |
|
252 |
|
253 return KErrNone; |
|
254 } |
|
255 |
|
256 |
|
257 TUint8* DWin32ShPool::Base() |
|
258 { |
|
259 return iWin32MemoryBase; |
|
260 } |
|
261 |
|
262 |
|
263 TUint8* DWin32ShPool::Base(DProcess* /*aProcess*/) |
|
264 { |
|
265 return iWin32MemoryBase; |
|
266 } |
|
267 |
|
268 |
|
269 TInt DWin32ShPool::AddToProcess(DProcess* aProcess, TUint aAttr) |
|
270 { |
|
271 __KTRACE_OPT(KEXEC, Kern::Printf("Adding DWin32ShPool %O to process %O", this, aProcess)); |
|
272 |
|
273 TInt r = KErrNone; |
|
274 |
|
275 Kern::MutexWait(*iProcessLock); |
|
276 LockPool(); |
|
277 DShPoolClient* client = reinterpret_cast<DShPoolClient*>(iClientMap->Find(reinterpret_cast<TUint>(aProcess))); |
|
278 UnlockPool(); |
|
279 |
|
280 if (!client) |
|
281 { |
|
282 client = new DShPoolClient; |
|
283 |
|
284 if (client) |
|
285 { |
|
286 client->iFlags = aAttr; |
|
287 r = iClientMap->Add(reinterpret_cast<TUint>(aProcess), client); |
|
288 |
|
289 if (r == KErrNone) |
|
290 { |
|
291 if (aProcess != K::TheKernelProcess) |
|
292 { |
|
293 r = aProcess->iHandles.Reserve(iTotalBuffers); |
|
294 |
|
295 if (r != KErrNone) |
|
296 { |
|
297 iClientMap->Remove(reinterpret_cast<TUint>(aProcess)); |
|
298 } |
|
299 } |
|
300 } |
|
301 |
|
302 if (r != KErrNone) |
|
303 { |
|
304 delete client; |
|
305 } |
|
306 } |
|
307 else |
|
308 { |
|
309 r = KErrNoMemory; |
|
310 } |
|
311 } |
|
312 else |
|
313 { |
|
314 LockPool(); |
|
315 client->iAccessCount++; |
|
316 UnlockPool(); |
|
317 } |
|
318 |
|
319 Kern::MutexSignal(*iProcessLock); |
|
320 |
|
321 return r; |
|
322 } |
|
323 |
|
324 |
|
325 TInt DWin32ShPool::DoCreate(TShPoolCreateInfo& aInfo) |
|
326 { |
|
327 TUint64 maxSize = static_cast<TUint64>(aInfo.iInfo.iMaxBufs) * static_cast<TUint64>(iBufGap); |
|
328 |
|
329 if (maxSize > static_cast<TUint64>(KMaxTInt)) |
|
330 { |
|
331 return KErrArgument; |
|
332 } |
|
333 |
|
334 __KTRACE_OPT(KMMU,Kern::Printf("DWin32ShPool::DoCreate (maxSize = 0x%08x, iBufGap = 0x%08x)", |
|
335 static_cast<TInt>(maxSize), iBufGap)); |
|
336 |
|
337 iWin32MemoryBase = (TUint8*) VirtualAlloc(NULL, (SIZE_T)maxSize, MEM_RESERVE, PAGE_READWRITE); |
|
338 if (iWin32MemoryBase == NULL) |
|
339 { |
|
340 return KErrNoMemory; |
|
341 } |
|
342 |
|
343 __KTRACE_OPT(KMMU,Kern::Printf("DWin32ShPool::DoCreate (iWin32MemoryBase = 0x%08x)", iWin32MemoryBase)); |
|
344 |
|
345 iBufMap = TBitMapAllocator::New(aInfo.iInfo.iMaxBufs, (TBool)ETrue); |
|
346 if (iBufMap == NULL) |
|
347 { |
|
348 return KErrNoMemory; |
|
349 } |
|
350 |
|
351 return KErrNone; |
|
352 } |
|
353 |
|
354 |
|
355 TBool DWin32ShPool::IsOpen(DProcess* /*aProcess*/) |
|
356 { |
|
357 // could do we some kind of check here? |
|
358 return (TBool)ETrue; |
|
359 } |
|
360 |
|
361 |
|
362 TInt DWin32ShPool::UpdateFreeList() |
|
363 { |
|
364 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32ShPool::UpdateFreeList")); |
|
365 |
|
366 SDblQue temp; |
|
367 SDblQueLink* anchor = reinterpret_cast<SDblQueLink*>(&iFreeList); |
|
368 |
|
369 LockPool(); |
|
370 while(!iAltFreeList.IsEmpty()) |
|
371 { |
|
372 // sort a temporary list of 'n' object with the lowest index first |
|
373 for (TInt n = 0; n < 8 && !iAltFreeList.IsEmpty(); ++n) |
|
374 { |
|
375 // bit of an assumption, lets assume that the lower indexes will be allocated and freed first |
|
376 // and therefore will be nearer the front of the list |
|
377 DShBuf* buf = _LOFF(iAltFreeList.GetFirst(), DShBuf, iObjLink); |
|
378 |
|
379 SDblQueLink* anchor = reinterpret_cast<SDblQueLink*>(&temp); |
|
380 SDblQueLink* pLink = temp.Last(); |
|
381 |
|
382 for (;;) |
|
383 { |
|
384 // traverse the list starting at the back |
|
385 if ((pLink != anchor) && (_LOFF(pLink, DShBuf, iObjLink)->iRelAddress > buf->iRelAddress)) |
|
386 { |
|
387 pLink = pLink->iPrev; |
|
388 } |
|
389 else |
|
390 { |
|
391 buf->iObjLink.InsertAfter(pLink); |
|
392 break; |
|
393 } |
|
394 } |
|
395 } |
|
396 |
|
397 // now merge with the free list |
|
398 while(!temp.IsEmpty()) |
|
399 { |
|
400 if (iFreeList.IsEmpty()) |
|
401 { |
|
402 iFreeList.MoveFrom(&temp); |
|
403 break; |
|
404 } |
|
405 |
|
406 // working backwards with the highest index |
|
407 DShBuf* buf = _LOFF(temp.Last(), DShBuf, iObjLink); |
|
408 SDblQueLink* pLink = iFreeList.Last(); |
|
409 |
|
410 while (!NKern::FMFlash(&iLock)) |
|
411 { |
|
412 if ((pLink != anchor) && (_LOFF(pLink, DShBuf, iObjLink)->iRelAddress > buf->iRelAddress)) |
|
413 { |
|
414 pLink = pLink->iPrev; |
|
415 } |
|
416 else |
|
417 { |
|
418 buf->iObjLink.Deque(); |
|
419 buf->iObjLink.InsertAfter(pLink); |
|
420 // next buffer |
|
421 if (temp.IsEmpty()) |
|
422 break; |
|
423 buf = _LOFF(temp.Last(), DShBuf, iObjLink); |
|
424 } |
|
425 } |
|
426 } |
|
427 NKern::FMFlash(&iLock); |
|
428 } |
|
429 UnlockPool(); |
|
430 |
|
431 __KTRACE_OPT(KMMU, Kern::Printf("<DWin32ShPool::UpdateFreeList")); |
|
432 return KErrNone; |
|
433 } |
|
434 |
|
435 |
|
436 void DWin32ShPool::Free(DShBuf* aBuf) |
|
437 { |
|
438 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32ShPool::Free (aBuf = 0x%08x, aBuf->Base() 0x%08x)", aBuf, aBuf->Base())); |
|
439 |
|
440 TLinAddr newAddr = (TLinAddr)aBuf->Base(); |
|
441 #ifdef _DEBUG |
|
442 memset((TAny*)newAddr,0xde,aBuf->Size()); |
|
443 #else |
|
444 memclr((TAny*)newAddr,aBuf->Size()); |
|
445 #endif |
|
446 |
|
447 LockPool(); |
|
448 #ifdef _DEBUG |
|
449 // Remove from allocated list |
|
450 aBuf->iObjLink.Deque(); |
|
451 #endif |
|
452 // we want to put the initial buffers at the head of the free list |
|
453 // and the grown buffers at the tail as this makes shrinking more efficient |
|
454 if (aBuf >= iInitialBuffersArray && aBuf < (iInitialBuffersArray + iInitialBuffers)) |
|
455 { |
|
456 iFreeList.AddHead(&aBuf->iObjLink); |
|
457 } |
|
458 else |
|
459 { |
|
460 iAltFreeList.Add(&aBuf->iObjLink); |
|
461 } |
|
462 |
|
463 ++iFreeBuffers; |
|
464 #ifdef _DEBUG |
|
465 --iAllocatedBuffers; |
|
466 #endif |
|
467 iPoolFlags &= ~EShPoolSuppressShrink; // Allow shrinking again, if it was blocked |
|
468 UnlockPool(); |
|
469 |
|
470 // queue ManagementDfc which completes notifications as appropriate |
|
471 if (HaveWorkToDo()) |
|
472 KickManagementDfc(); |
|
473 |
|
474 Close(NULL); // decrement pool reference count |
|
475 } |
|
476 |
|
477 // Kernel side API |
|
478 TInt DWin32ShPool::Alloc(DShBuf*& aShBuf) |
|
479 { |
|
480 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32ShPool::Alloc (DShBuf)")); |
|
481 |
|
482 TInt r = KErrNoMemory; |
|
483 aShBuf = NULL; |
|
484 |
|
485 LockPool(); |
|
486 |
|
487 if (!iFreeList.IsEmpty()) |
|
488 { |
|
489 aShBuf = _LOFF(iFreeList.GetFirst(), DShBuf, iObjLink); |
|
490 #ifdef _DEBUG |
|
491 iAllocated.Add(&aShBuf->iObjLink); |
|
492 iAllocatedBuffers++; |
|
493 #endif |
|
494 --iFreeBuffers; |
|
495 Open(); // increment pool reference count |
|
496 r = KErrNone; |
|
497 } |
|
498 else |
|
499 { |
|
500 // try alternative free list |
|
501 if (!iAltFreeList.IsEmpty()) |
|
502 { |
|
503 aShBuf = _LOFF(iAltFreeList.GetFirst(), DShBuf, iObjLink); |
|
504 #ifdef _DEBUG |
|
505 iAllocated.Add(&aShBuf->iObjLink); |
|
506 iAllocatedBuffers++; |
|
507 #endif |
|
508 --iFreeBuffers; |
|
509 Open(); // increment pool reference count |
|
510 r = KErrNone; |
|
511 } |
|
512 } |
|
513 |
|
514 UnlockPool(); |
|
515 |
|
516 if (HaveWorkToDo()) |
|
517 KickManagementDfc(); |
|
518 |
|
519 __KTRACE_OPT(KMMU, Kern::Printf("<DWin32ShPool::Alloc return buf = 0x%08x", aShBuf)); |
|
520 return r; |
|
521 } |
|
522 |
|
523 |
|
524 DWin32AlignedShPool::DWin32AlignedShPool() |
|
525 : DWin32ShPool() |
|
526 { |
|
527 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32AlignedShPool::DWin32AlignedShPool")); |
|
528 } |
|
529 |
|
530 |
|
531 DWin32AlignedShPool::~DWin32AlignedShPool() |
|
532 { |
|
533 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32AlignedShPool::~DWin32AlignedShPool")); |
|
534 } |
|
535 |
|
536 |
|
537 TInt DWin32AlignedShPool::DoCreate(TShPoolCreateInfo& aInfo) |
|
538 { |
|
539 TInt r; |
|
540 // Create Chunk |
|
541 r = DWin32ShPool::DoCreate(aInfo); |
|
542 if (r != KErrNone) |
|
543 { |
|
544 return r; |
|
545 } |
|
546 |
|
547 if (iPoolFlags & EShPoolGuardPages) |
|
548 { |
|
549 TUint numOfBytes = iBufGap - MM::RamPageSize; |
|
550 iCommittedPages = MM::RoundToPageSize(iInitialBuffers * numOfBytes) >> MM::RamPageShift; |
|
551 |
|
552 for (TUint i = 0; i < iInitialBuffers; ++i) |
|
553 { |
|
554 TUint offset = iBufGap * i; |
|
555 |
|
556 MM::Wait(); |
|
557 if (MM::Commit(reinterpret_cast<TLinAddr>(iWin32MemoryBase+offset), numOfBytes, 0xFF, EFalse) != KErrNone) |
|
558 { |
|
559 MM::Signal(); |
|
560 return KErrNoMemory; |
|
561 } |
|
562 iWin32MemorySize += numOfBytes; |
|
563 |
|
564 MM::Signal(); |
|
565 } |
|
566 |
|
567 iMaxPages = MM::RoundToPageSize(aInfo.iInfo.iMaxBufs * numOfBytes) >> MM::RamPageShift; |
|
568 } |
|
569 else |
|
570 { |
|
571 // Make sure we give the caller the number of buffers they were expecting |
|
572 iCommittedPages = MM::RoundToPageSize(iInitialBuffers * iBufGap) >> MM::RamPageShift; |
|
573 MM::Wait(); |
|
574 if (MM::Commit(reinterpret_cast<TLinAddr>(iWin32MemoryBase), iCommittedPages << MM::RamPageShift, 0xFF, EFalse) != KErrNone) |
|
575 { |
|
576 MM::Signal(); |
|
577 return KErrNoMemory; |
|
578 } |
|
579 iWin32MemorySize = iCommittedPages << MM::RamPageShift; |
|
580 |
|
581 MM::Signal(); |
|
582 |
|
583 iMaxPages = MM::RoundToPageSize(aInfo.iInfo.iMaxBufs * iBufGap) >> MM::RamPageShift; |
|
584 } |
|
585 |
|
586 return r; |
|
587 } |
|
588 |
|
589 |
|
590 TInt DWin32AlignedShPool::SetBufferWindow(DProcess* /*aProcess*/, TInt /*aWindowSize*/ ) |
|
591 { |
|
592 return KErrNone; |
|
593 } |
|
594 |
|
595 |
|
596 TInt DWin32AlignedShPool::GrowPool() |
|
597 { |
|
598 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32AlignedShPool::GrowPool()")); |
|
599 |
|
600 Kern::MutexWait(*iProcessLock); |
|
601 |
|
602 // How many bytes to commit for each new buffer (must be whole number of pages) |
|
603 TUint bytes = (iPoolFlags & EShPoolGuardPages) ? iBufGap - MM::RamPageSize : iBufGap; |
|
604 |
|
605 __ASSERT_DEBUG(!(bytes % MM::RamPageSize), Kern::PanicCurrentThread(KLitDWin32AlignedShPool, __LINE__)); |
|
606 |
|
607 TInt pages = bytes >> MM::RamPageShift; |
|
608 |
|
609 TUint32 headroom = iMaxBuffers - iTotalBuffers; |
|
610 |
|
611 // How many buffers to grow by? |
|
612 TUint32 grow = mult_fx248(iTotalBuffers, iGrowByRatio); |
|
613 if (grow == 0) // Handle round-to-zero |
|
614 grow = 1; |
|
615 if (grow > headroom) |
|
616 grow = headroom; |
|
617 |
|
618 TInt r = KErrNone; |
|
619 SDblQue temp; |
|
620 |
|
621 TUint i; |
|
622 for (i = 0; i < grow; ++i) |
|
623 { |
|
624 TInt offset = iBufMap->Alloc(); |
|
625 |
|
626 if (offset < 0) |
|
627 { |
|
628 r = KErrNoMemory; |
|
629 break; |
|
630 } |
|
631 |
|
632 offset *= iBufGap; |
|
633 |
|
634 MM::Wait(); |
|
635 if (MM::Commit(reinterpret_cast<TLinAddr>(iWin32MemoryBase+offset), bytes, 0xFF, EFalse) != KErrNone) |
|
636 { |
|
637 r = KErrNoMemory; |
|
638 } |
|
639 iWin32MemorySize += bytes; |
|
640 MM::Signal(); |
|
641 |
|
642 if (r != KErrNone) |
|
643 { |
|
644 iBufMap->Free(offset / iBufGap); |
|
645 break; |
|
646 } |
|
647 |
|
648 DWin32ShBuf *buf = new DWin32ShBuf(this, offset); |
|
649 |
|
650 if (buf == NULL) |
|
651 { |
|
652 MM::Wait(); |
|
653 MM::Decommit(reinterpret_cast<TLinAddr>(iWin32MemoryBase+offset), bytes); |
|
654 iWin32MemorySize -= bytes; |
|
655 MM::Signal(); |
|
656 iBufMap->Free(offset / iBufGap); |
|
657 r = KErrNoMemory; |
|
658 break; |
|
659 } |
|
660 |
|
661 TInt r = buf->Construct(); |
|
662 |
|
663 if (r != KErrNone) |
|
664 { |
|
665 MM::Wait(); |
|
666 MM::Decommit(reinterpret_cast<TLinAddr>(iWin32MemoryBase+offset), bytes); |
|
667 iWin32MemorySize -= bytes; |
|
668 MM::Signal(); |
|
669 iBufMap->Free(offset / iBufGap); |
|
670 buf->DObject::Close(NULL); |
|
671 break; |
|
672 } |
|
673 |
|
674 iCommittedPages += pages; |
|
675 |
|
676 temp.Add(&buf->iObjLink); |
|
677 } |
|
678 |
|
679 r = UpdateReservedHandles(i); |
|
680 |
|
681 if (r == KErrNone) |
|
682 { |
|
683 LockPool(); |
|
684 iFreeList.MoveFrom(&temp); |
|
685 iFreeBuffers += i; |
|
686 iTotalBuffers += i; |
|
687 UnlockPool(); |
|
688 } |
|
689 else |
|
690 { |
|
691 // else delete buffers |
|
692 SDblQueLink *pLink; |
|
693 while ((pLink = temp.GetFirst()) != NULL) |
|
694 { |
|
695 DShBuf* buf = _LOFF(pLink, DShBuf, iObjLink); |
|
696 TLinAddr offset = buf->iRelAddress; |
|
697 iBufMap->Free(offset / iBufGap); |
|
698 MM::Wait(); |
|
699 MM::Decommit(reinterpret_cast<TLinAddr>(iWin32MemoryBase+offset), bytes); |
|
700 iWin32MemorySize -= bytes; |
|
701 MM::Signal(); |
|
702 iCommittedPages -= pages; |
|
703 buf->DObject::Close(NULL); |
|
704 } |
|
705 } |
|
706 |
|
707 CalculateGrowShrinkTriggers(); |
|
708 |
|
709 Kern::MutexSignal(*iProcessLock); |
|
710 |
|
711 __KTRACE_OPT(KMMU, Kern::Printf("<DWin32AlignedShPool::GrowPool()")); |
|
712 return r; |
|
713 } // DWin32AlignedShPool::GrowPool |
|
714 |
|
715 |
|
716 TInt DWin32AlignedShPool::ShrinkPool() |
|
717 { |
|
718 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32AlignedShPool::ShrinkPool()")); |
|
719 |
|
720 Kern::MutexWait(*iProcessLock); |
|
721 |
|
722 // How many bytes to commit for each new buffer (must be whole number of pages) |
|
723 TUint bytes = (iPoolFlags & EShPoolGuardPages) ? iBufGap - MM::RamPageSize : iBufGap; |
|
724 |
|
725 __ASSERT_DEBUG(!(bytes % MM::RamPageSize), Kern::PanicCurrentThread(KLitDWin32AlignedShPool, __LINE__)); |
|
726 |
|
727 TInt pages = bytes >> MM::RamPageShift; |
|
728 |
|
729 // Grab pool stats |
|
730 TUint32 grownBy = iTotalBuffers - iInitialBuffers; |
|
731 |
|
732 // How many buffers to shrink by? |
|
733 TUint32 shrink = mult_fx248(iTotalBuffers, iShrinkByRatio); |
|
734 if (shrink == 0) // Handle round-to-zero |
|
735 shrink = 1; |
|
736 if (shrink > grownBy) |
|
737 shrink = grownBy; |
|
738 if (shrink > iFreeBuffers) |
|
739 shrink = iFreeBuffers; |
|
740 |
|
741 // work backwards |
|
742 TUint i; |
|
743 for (i = 0; i < shrink; ++i) |
|
744 { |
|
745 LockPool(); |
|
746 if (iFreeList.IsEmpty()) |
|
747 { |
|
748 UnlockPool(); |
|
749 break; |
|
750 } |
|
751 // work from the back of the queue |
|
752 SDblQueLink *pLink = iFreeList.Last(); |
|
753 |
|
754 DShBuf* pBuf = _LOFF(pLink, DShBuf, iObjLink); |
|
755 |
|
756 if (pBuf >= iInitialBuffersArray && pBuf < (iInitialBuffersArray + iInitialBuffers)) |
|
757 { |
|
758 UnlockPool(); |
|
759 break; |
|
760 } |
|
761 |
|
762 --iFreeBuffers; |
|
763 --iTotalBuffers; |
|
764 pLink->Deque(); |
|
765 iCommittedPages -= pages; |
|
766 UnlockPool(); |
|
767 |
|
768 TLinAddr offset = pBuf->iRelAddress; |
|
769 |
|
770 iBufMap->Free(offset / iBufGap); |
|
771 |
|
772 MM::Wait(); |
|
773 MM::Decommit(reinterpret_cast<TLinAddr>(iWin32MemoryBase+offset), iBufSize); |
|
774 iWin32MemorySize -= iBufSize; |
|
775 MM::Signal(); |
|
776 pBuf->DObject::Close(NULL); |
|
777 } |
|
778 |
|
779 TInt r = UpdateReservedHandles(-(TInt)i); |
|
780 |
|
781 // If we couldn't shrink the pool by this many buffers, wait until we Free() another |
|
782 // buffer before trying to shrink again. |
|
783 if (i < shrink) |
|
784 iPoolFlags |= EShPoolSuppressShrink; |
|
785 |
|
786 CalculateGrowShrinkTriggers(); |
|
787 |
|
788 Kern::MutexSignal(*iProcessLock); |
|
789 |
|
790 __KTRACE_OPT(KMMU, Kern::Printf("<DWin32AlignedShPool::ShrinkPool()")); |
|
791 return r; |
|
792 } // DWin32AlignedShPool::ShrinkPool |
|
793 |
|
794 |
|
795 DWin32NonAlignedShPool::DWin32NonAlignedShPool() |
|
796 : DWin32ShPool() |
|
797 { |
|
798 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32NonAlignedShPool::DWin32NonAlignedShPool")); |
|
799 } |
|
800 |
|
801 |
|
802 DWin32NonAlignedShPool::~DWin32NonAlignedShPool() |
|
803 { |
|
804 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32NonAlignedShPool::~DWin32NonAlignedShPool")); |
|
805 |
|
806 delete iPagesMap; |
|
807 } |
|
808 |
|
809 |
|
810 TInt DWin32NonAlignedShPool::DoCreate(TShPoolCreateInfo& aInfo) |
|
811 { |
|
812 // Create Chunk |
|
813 TInt r; |
|
814 |
|
815 r = DWin32ShPool::DoCreate(aInfo); |
|
816 |
|
817 if (r != KErrNone) |
|
818 { |
|
819 return r; |
|
820 } |
|
821 |
|
822 if (iPoolFlags & EShPoolPhysicalMemoryPool) |
|
823 { |
|
824 return KErrNotSupported; |
|
825 } |
|
826 else |
|
827 { |
|
828 // Make sure we give the caller the number of buffers they were expecting |
|
829 iCommittedPages = MM::RoundToPageSize(iInitialBuffers * iBufGap) >> MM::RamPageShift; |
|
830 |
|
831 MM::Wait(); |
|
832 if (MM::Commit(reinterpret_cast<TLinAddr>(iWin32MemoryBase), iCommittedPages << MM::RamPageShift, 0xFF, EFalse) != KErrNone) |
|
833 { |
|
834 MM::Signal(); |
|
835 return KErrNoMemory; |
|
836 } |
|
837 iWin32MemorySize = iCommittedPages << MM::RamPageShift; |
|
838 |
|
839 MM::Signal(); |
|
840 iMaxPages = MM::RoundToPageSize(aInfo.iInfo.iMaxBufs * iBufGap) >> MM::RamPageShift; |
|
841 } |
|
842 |
|
843 iPagesMap = TBitMapAllocator::New(iMaxPages, (TBool)ETrue); |
|
844 |
|
845 if(!iPagesMap) |
|
846 { |
|
847 return KErrNoMemory; |
|
848 } |
|
849 |
|
850 iPagesMap->Alloc(0, iCommittedPages); |
|
851 return r; |
|
852 } |
|
853 |
|
854 |
|
855 void DWin32NonAlignedShPool::FreeBufferPages(TUint aOffset) |
|
856 { |
|
857 TLinAddr firstByte = aOffset; // offset of first byte in buffer |
|
858 TLinAddr lastByte = firstByte+iBufGap-1; // offset of last byte in buffer |
|
859 TUint firstPage = firstByte>>MM::RamPageShift; // index of first page containing part of the buffer |
|
860 TUint lastPage = lastByte>>MM::RamPageShift; // index of last page containing part of the buffer |
|
861 |
|
862 TUint firstBuffer = (firstByte&~(MM::RamPageSize - 1))/iBufGap; // index of first buffer which lies in firstPage |
|
863 TUint lastBuffer = (lastByte|(MM::RamPageSize - 1))/iBufGap; // index of last buffer which lies in lastPage |
|
864 TUint thisBuffer = firstByte/iBufGap; // index of the buffer to be freed |
|
865 |
|
866 // Ensure lastBuffer is within bounds (there may be room in the last |
|
867 // page for more buffers than we have allocated). |
|
868 if (lastBuffer >= iMaxBuffers) |
|
869 lastBuffer = iMaxBuffers-1; |
|
870 |
|
871 if(firstBuffer!=thisBuffer && iBufMap->NotFree(firstBuffer,thisBuffer-firstBuffer)) |
|
872 { |
|
873 // first page has other allocated buffers in it, |
|
874 // so we can't free it and must move on to next one... |
|
875 if (firstPage >= lastPage) |
|
876 return; |
|
877 ++firstPage; |
|
878 } |
|
879 |
|
880 if(lastBuffer!=thisBuffer && iBufMap->NotFree(thisBuffer+1,lastBuffer-thisBuffer)) |
|
881 { |
|
882 // last page has other allocated buffers in it, |
|
883 // so we can't free it and must step back to previous one... |
|
884 if (lastPage <= firstPage) |
|
885 return; |
|
886 --lastPage; |
|
887 } |
|
888 |
|
889 if(firstPage<=lastPage) |
|
890 { |
|
891 // we can free pages firstPage trough to lastPage... |
|
892 TUint numPages = lastPage-firstPage+1; |
|
893 iPagesMap->SelectiveFree(firstPage,numPages); |
|
894 MM::Wait(); |
|
895 MM::Decommit(reinterpret_cast<TLinAddr>(iWin32MemoryBase+(firstPage << MM::RamPageShift)), (numPages << MM::RamPageShift)); |
|
896 iWin32MemorySize -= (numPages << MM::RamPageShift); |
|
897 MM::Signal(); |
|
898 iCommittedPages -= numPages; |
|
899 } |
|
900 } |
|
901 |
|
902 |
|
903 TInt DWin32NonAlignedShPool::GrowPool() |
|
904 { |
|
905 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32NonAlignedShPool::GrowPool()")); |
|
906 |
|
907 Kern::MutexWait(*iProcessLock); |
|
908 |
|
909 TUint32 headroom = iMaxBuffers - iTotalBuffers; |
|
910 |
|
911 // How many buffers to grow by? |
|
912 TUint32 grow = mult_fx248(iTotalBuffers, iGrowByRatio); |
|
913 if (grow == 0) // Handle round-to-zero |
|
914 grow = 1; |
|
915 if (grow > headroom) |
|
916 grow = headroom; |
|
917 |
|
918 TInt r = KErrNone; |
|
919 SDblQue temp; |
|
920 |
|
921 TUint i; |
|
922 for (i = 0; i < grow; ++i) |
|
923 { |
|
924 TInt offset = iBufMap->Alloc(); |
|
925 |
|
926 if (offset < 0) |
|
927 { |
|
928 r = KErrNoMemory; |
|
929 break; |
|
930 } |
|
931 |
|
932 offset *= iBufGap; |
|
933 |
|
934 TInt lastPage = (offset + iBufSize - 1) >> MM::RamPageShift; |
|
935 |
|
936 // Allocate one page at a time. |
|
937 for (TInt page = offset >> MM::RamPageShift; page <= lastPage; ++page) |
|
938 { |
|
939 // Is the page allocated? |
|
940 if (iPagesMap->NotAllocated(page, 1)) |
|
941 { |
|
942 MM::Wait(); |
|
943 if (MM::Commit(reinterpret_cast<TLinAddr>(iWin32MemoryBase+(page << MM::RamPageShift)), MM::RamPageSize, 0xFF, EFalse) != KErrNone) |
|
944 { |
|
945 MM::Signal(); |
|
946 r = KErrNoMemory; |
|
947 break; |
|
948 } |
|
949 iWin32MemorySize += MM::RamPageSize; |
|
950 |
|
951 MM::Signal(); |
|
952 ++iCommittedPages; |
|
953 iPagesMap->Alloc(page, 1); |
|
954 } |
|
955 } |
|
956 |
|
957 if (r != KErrNone) |
|
958 { |
|
959 iBufMap->Free(offset / iBufGap); |
|
960 FreeBufferPages(offset); |
|
961 break; |
|
962 } |
|
963 |
|
964 DWin32ShBuf *buf = new DWin32ShBuf(this, offset); |
|
965 |
|
966 if (buf == NULL) |
|
967 { |
|
968 iBufMap->Free(offset / iBufGap); |
|
969 FreeBufferPages(offset); |
|
970 r = KErrNoMemory; |
|
971 break; |
|
972 } |
|
973 |
|
974 r = buf->Construct(); |
|
975 |
|
976 if (r != KErrNone) |
|
977 { |
|
978 iBufMap->Free(offset / iBufGap); |
|
979 FreeBufferPages(offset); |
|
980 buf->DObject::Close(NULL); |
|
981 break; |
|
982 } |
|
983 |
|
984 temp.Add(&buf->iObjLink); |
|
985 } |
|
986 |
|
987 r = UpdateReservedHandles(i); |
|
988 |
|
989 if (r == KErrNone) |
|
990 { |
|
991 LockPool(); |
|
992 iFreeList.MoveFrom(&temp); |
|
993 iFreeBuffers += i; |
|
994 iTotalBuffers += i; |
|
995 UnlockPool(); |
|
996 } |
|
997 else |
|
998 { |
|
999 // couldn't reserve handles so have no choice but to |
|
1000 // delete the buffers |
|
1001 __KTRACE_OPT(KMMU, Kern::Printf("GrowPool failed with %d, deleting buffers", r)); |
|
1002 SDblQueLink *pLink; |
|
1003 while ((pLink = temp.GetFirst()) != NULL) |
|
1004 { |
|
1005 DShBuf* buf = _LOFF(pLink, DShBuf, iObjLink); |
|
1006 TLinAddr offset = buf->iRelAddress; |
|
1007 iBufMap->Free(offset / iBufGap); |
|
1008 FreeBufferPages(offset); |
|
1009 buf->DObject::Close(NULL); |
|
1010 } |
|
1011 __KTRACE_OPT(KMMU, Kern::Printf("Buffers deleted")); |
|
1012 } |
|
1013 |
|
1014 CalculateGrowShrinkTriggers(); |
|
1015 |
|
1016 Kern::MutexSignal(*iProcessLock); |
|
1017 |
|
1018 __KTRACE_OPT(KMMU, Kern::Printf("<DWin32NonAlignedShPool::GrowPool()")); |
|
1019 return r; |
|
1020 } // DWin32NonAlignedShPool::GrowPool |
|
1021 |
|
1022 |
|
1023 TInt DWin32NonAlignedShPool::ShrinkPool() |
|
1024 { |
|
1025 __KTRACE_OPT(KMMU, Kern::Printf(">DWin32NonAlignedShPool::ShrinkPool()")); |
|
1026 |
|
1027 Kern::MutexWait(*iProcessLock); |
|
1028 |
|
1029 // Grab pool stats |
|
1030 TUint32 grownBy = iTotalBuffers - iInitialBuffers; |
|
1031 |
|
1032 // How many buffers to shrink by? |
|
1033 TUint32 shrink = mult_fx248(iTotalBuffers, iShrinkByRatio); |
|
1034 if (shrink == 0) // Handle round-to-zero |
|
1035 shrink = 1; |
|
1036 if (shrink > grownBy) |
|
1037 shrink = grownBy; |
|
1038 if (shrink > iFreeBuffers) |
|
1039 shrink = iFreeBuffers; |
|
1040 |
|
1041 TUint i; |
|
1042 for (i = 0; i < shrink; ++i) |
|
1043 { |
|
1044 LockPool(); |
|
1045 if (iFreeList.IsEmpty()) |
|
1046 { |
|
1047 UnlockPool(); |
|
1048 break; |
|
1049 } |
|
1050 // work from the back of the queue |
|
1051 SDblQueLink *pLink = iFreeList.Last(); |
|
1052 |
|
1053 DShBuf* pBuf = _LOFF(pLink, DShBuf, iObjLink); |
|
1054 |
|
1055 if (pBuf >= iInitialBuffersArray && pBuf < (iInitialBuffersArray + iInitialBuffers)) |
|
1056 { |
|
1057 UnlockPool(); |
|
1058 break; |
|
1059 } |
|
1060 |
|
1061 --iFreeBuffers; |
|
1062 --iTotalBuffers; |
|
1063 pLink->Deque(); |
|
1064 UnlockPool(); |
|
1065 |
|
1066 TLinAddr offset = pBuf->iRelAddress; |
|
1067 |
|
1068 iBufMap->Free(offset / iBufGap); |
|
1069 FreeBufferPages(offset); |
|
1070 pBuf->DObject::Close(NULL); |
|
1071 } |
|
1072 |
|
1073 UpdateReservedHandles(-(TInt)i); |
|
1074 |
|
1075 // If we couldn't shrink the pool by this many buffers, wait until we Free() another |
|
1076 // buffer before trying to shrink again. |
|
1077 if (i < shrink) |
|
1078 iPoolFlags |= EShPoolSuppressShrink; |
|
1079 |
|
1080 CalculateGrowShrinkTriggers(); |
|
1081 |
|
1082 Kern::MutexSignal(*iProcessLock); |
|
1083 |
|
1084 __KTRACE_OPT(KMMU, Kern::Printf("<DWin32NonAlignedShPool::ShrinkPool()")); |
|
1085 |
|
1086 return KErrNone; |
|
1087 } // DWin32NonAlignedShPool::ShrinkPool |